[ 465.726877] env[63293]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=63293) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.727199] env[63293]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=63293) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.727309] env[63293]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=63293) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.727600] env[63293]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 465.819373] env[63293]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=63293) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 465.829314] env[63293]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=63293) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 466.430190] env[63293]: INFO nova.virt.driver [None req-239d82ca-1e7b-473f-83ac-5984e77cd6c0 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 466.502953] env[63293]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 466.503194] env[63293]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 466.503270] env[63293]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=63293) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 469.601175] env[63293]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-18fd5983-132b-41a0-b90c-4310fc53a9c8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.616804] env[63293]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=63293) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 469.616971] env[63293]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-8984842e-96d7-46bf-8915-e90f76a1899a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.650144] env[63293]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 40546. [ 469.650318] env[63293]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.147s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 469.650790] env[63293]: INFO nova.virt.vmwareapi.driver [None req-239d82ca-1e7b-473f-83ac-5984e77cd6c0 None None] VMware vCenter version: 7.0.3 [ 469.654182] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40aa63d-4fe4-4574-bbb0-cc469b6420e9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.671653] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-758d526b-728f-4100-ae10-37d2810f65be {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.677531] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fcf5952-841c-4b7d-9abd-8477a3b9446a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.683950] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a988026d-fbce-4542-b922-f78d82afb389 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.697157] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83b306c-5d16-4a59-9114-d5a2970dc743 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.703249] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50be4799-ef09-4a94-b74b-6e98f5a06519 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.735382] env[63293]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-6e022625-7191-44ce-87f2-299b8e444f78 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.740355] env[63293]: DEBUG nova.virt.vmwareapi.driver [None req-239d82ca-1e7b-473f-83ac-5984e77cd6c0 None None] Extension org.openstack.compute already exists. {{(pid=63293) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 469.742924] env[63293]: INFO nova.compute.provider_config [None req-239d82ca-1e7b-473f-83ac-5984e77cd6c0 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 470.249018] env[63293]: DEBUG nova.context [None req-239d82ca-1e7b-473f-83ac-5984e77cd6c0 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),c1d24a51-5058-4b9e-b041-dc548cfa9d96(cell1) {{(pid=63293) load_cells /opt/stack/nova/nova/context.py:464}} [ 470.249018] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 470.249281] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 470.249847] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.250290] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Acquiring lock "c1d24a51-5058-4b9e-b041-dc548cfa9d96" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 470.250470] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Lock "c1d24a51-5058-4b9e-b041-dc548cfa9d96" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 470.251539] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Lock "c1d24a51-5058-4b9e-b041-dc548cfa9d96" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.272240] env[63293]: INFO dbcounter [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Registered counter for database nova_cell0 [ 470.280327] env[63293]: INFO dbcounter [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Registered counter for database nova_cell1 [ 470.283669] env[63293]: DEBUG oslo_db.sqlalchemy.engines [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63293) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 470.284041] env[63293]: DEBUG oslo_db.sqlalchemy.engines [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63293) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 470.288699] env[63293]: ERROR nova.db.main.api [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 470.288699] env[63293]: result = function(*args, **kwargs) [ 470.288699] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 470.288699] env[63293]: return func(*args, **kwargs) [ 470.288699] env[63293]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 470.288699] env[63293]: result = fn(*args, **kwargs) [ 470.288699] env[63293]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 470.288699] env[63293]: return f(*args, **kwargs) [ 470.288699] env[63293]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 470.288699] env[63293]: return db.service_get_minimum_version(context, binaries) [ 470.288699] env[63293]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 470.288699] env[63293]: _check_db_access() [ 470.288699] env[63293]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 470.288699] env[63293]: stacktrace = ''.join(traceback.format_stack()) [ 470.288699] env[63293]: [ 470.289696] env[63293]: ERROR nova.db.main.api [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 470.289696] env[63293]: result = function(*args, **kwargs) [ 470.289696] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 470.289696] env[63293]: return func(*args, **kwargs) [ 470.289696] env[63293]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 470.289696] env[63293]: result = fn(*args, **kwargs) [ 470.289696] env[63293]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 470.289696] env[63293]: return f(*args, **kwargs) [ 470.289696] env[63293]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 470.289696] env[63293]: return db.service_get_minimum_version(context, binaries) [ 470.289696] env[63293]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 470.289696] env[63293]: _check_db_access() [ 470.289696] env[63293]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 470.289696] env[63293]: stacktrace = ''.join(traceback.format_stack()) [ 470.289696] env[63293]: [ 470.290245] env[63293]: WARNING nova.objects.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 470.290245] env[63293]: WARNING nova.objects.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Failed to get minimum service version for cell c1d24a51-5058-4b9e-b041-dc548cfa9d96 [ 470.290603] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Acquiring lock "singleton_lock" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 470.290759] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Acquired lock "singleton_lock" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 470.290996] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Releasing lock "singleton_lock" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 470.291352] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Full set of CONF: {{(pid=63293) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 470.291503] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ******************************************************************************** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 470.291630] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Configuration options gathered from: {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 470.291768] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 470.291960] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 470.292112] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ================================================================================ {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 470.292326] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] allow_resize_to_same_host = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.292493] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] arq_binding_timeout = 300 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.292622] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] backdoor_port = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.292746] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] backdoor_socket = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.292906] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] block_device_allocate_retries = 60 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.293082] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] block_device_allocate_retries_interval = 3 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.293254] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cert = self.pem {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.293418] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.293584] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute_monitors = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.293748] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] config_dir = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.293911] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] config_drive_format = iso9660 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.294049] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.294214] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] config_source = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.294403] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] console_host = devstack {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.294575] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] control_exchange = nova {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.294735] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cpu_allocation_ratio = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.294895] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] daemon = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.295067] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] debug = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.295241] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] default_access_ip_network_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.295412] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] default_availability_zone = nova {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.295566] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] default_ephemeral_format = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.295726] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] default_green_pool_size = 1000 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.295962] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.296156] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] default_schedule_zone = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.296339] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] disk_allocation_ratio = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.296504] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] enable_new_services = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.296683] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] enabled_apis = ['osapi_compute'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.296846] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] enabled_ssl_apis = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.297012] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] flat_injected = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.297177] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] force_config_drive = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.297351] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] force_raw_images = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.297539] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] graceful_shutdown_timeout = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.297698] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] heal_instance_info_cache_interval = 60 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.297903] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] host = cpu-1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.298087] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] initial_cpu_allocation_ratio = 4.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.298251] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] initial_disk_allocation_ratio = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.298409] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] initial_ram_allocation_ratio = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.298614] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.298775] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] instance_build_timeout = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.298934] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] instance_delete_interval = 300 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.299108] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] instance_format = [instance: %(uuid)s] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.299273] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] instance_name_template = instance-%08x {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.299431] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] instance_usage_audit = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.299598] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] instance_usage_audit_period = month {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.299761] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.299921] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] instances_path = /opt/stack/data/nova/instances {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.300108] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] internal_service_availability_zone = internal {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.300270] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] key = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.300451] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] live_migration_retry_count = 30 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.300628] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] log_color = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.300788] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] log_config_append = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.300954] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.301127] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] log_dir = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.301281] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] log_file = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.301407] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] log_options = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.301563] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] log_rotate_interval = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.301729] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] log_rotate_interval_type = days {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.301892] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] log_rotation_type = none {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.302026] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.302154] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.302319] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.302488] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.302616] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.302775] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] long_rpc_timeout = 1800 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.302928] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] max_concurrent_builds = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.303094] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] max_concurrent_live_migrations = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.303273] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] max_concurrent_snapshots = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.303441] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] max_local_block_devices = 3 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.303596] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] max_logfile_count = 30 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.303752] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] max_logfile_size_mb = 200 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.303908] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] maximum_instance_delete_attempts = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.304098] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] metadata_listen = 0.0.0.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.304272] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] metadata_listen_port = 8775 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.304439] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] metadata_workers = 2 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.304597] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] migrate_max_retries = -1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.304761] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] mkisofs_cmd = genisoimage {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.304962] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] my_block_storage_ip = 10.180.1.21 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.305105] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] my_ip = 10.180.1.21 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.305294] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] network_allocate_retries = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.305468] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.305655] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] osapi_compute_listen = 0.0.0.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.305818] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] osapi_compute_listen_port = 8774 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.305999] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] osapi_compute_unique_server_name_scope = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.306182] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] osapi_compute_workers = 2 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.306370] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] password_length = 12 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.306534] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] periodic_enable = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.306692] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] periodic_fuzzy_delay = 60 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.306858] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] pointer_model = usbtablet {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.307028] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] preallocate_images = none {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.307190] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] publish_errors = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.307320] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] pybasedir = /opt/stack/nova {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.307473] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ram_allocation_ratio = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.307628] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] rate_limit_burst = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.307792] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] rate_limit_except_level = CRITICAL {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.307949] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] rate_limit_interval = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.308116] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] reboot_timeout = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.308272] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] reclaim_instance_interval = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.308424] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] record = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.308587] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] reimage_timeout_per_gb = 60 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.308747] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] report_interval = 120 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.308903] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] rescue_timeout = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.309077] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] reserved_host_cpus = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.309251] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] reserved_host_disk_mb = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.309422] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] reserved_host_memory_mb = 512 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.309578] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] reserved_huge_pages = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.309734] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] resize_confirm_window = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.309887] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] resize_fs_using_block_device = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.310049] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] resume_guests_state_on_host_boot = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.310219] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.310376] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] rpc_response_timeout = 60 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.310529] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] run_external_periodic_tasks = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.310692] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] running_deleted_instance_action = reap {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.310850] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] running_deleted_instance_poll_interval = 1800 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.311014] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] running_deleted_instance_timeout = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.311176] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] scheduler_instance_sync_interval = 120 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.311342] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] service_down_time = 720 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.311503] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] servicegroup_driver = db {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.311656] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] shell_completion = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.311811] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] shelved_offload_time = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.311967] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] shelved_poll_interval = 3600 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.312141] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] shutdown_timeout = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.312340] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] source_is_ipv6 = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.312538] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ssl_only = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.312784] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.312952] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] sync_power_state_interval = 600 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.313139] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] sync_power_state_pool_size = 1000 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.313311] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] syslog_log_facility = LOG_USER {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.313468] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] tempdir = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.313626] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] timeout_nbd = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.313791] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] transport_url = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.313948] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] update_resources_interval = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.314115] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] use_cow_images = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.314272] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] use_eventlog = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.314429] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] use_journal = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.314586] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] use_json = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.314739] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] use_rootwrap_daemon = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.314892] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] use_stderr = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.315053] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] use_syslog = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.315242] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vcpu_pin_set = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.315444] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plugging_is_fatal = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.315614] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plugging_timeout = 300 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.315777] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] virt_mkfs = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.315935] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] volume_usage_poll_interval = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.316103] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] watch_log_file = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.316280] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] web = /usr/share/spice-html5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.316470] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.316636] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.316798] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_brick.wait_mpath_device_interval = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.316976] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_concurrency.disable_process_locking = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.317523] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.317719] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.317893] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.318090] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_metrics.metrics_process_name = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.318275] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.318454] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.318628] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.auth_strategy = keystone {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.318794] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.compute_link_prefix = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.318970] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.319156] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.dhcp_domain = novalocal {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.319326] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.enable_instance_password = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.319493] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.glance_link_prefix = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.319659] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.319831] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.instance_list_cells_batch_strategy = distributed {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.319993] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.instance_list_per_project_cells = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.320168] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.list_records_by_skipping_down_cells = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.320332] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.local_metadata_per_cell = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.320499] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.max_limit = 1000 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.320666] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.metadata_cache_expiration = 15 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.320837] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.neutron_default_tenant_id = default {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.321013] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.response_validation = warn {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.321189] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.use_neutron_default_nets = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.321416] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.321685] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.vendordata_dynamic_failure_fatal = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.321958] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.322270] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.vendordata_dynamic_ssl_certfile = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.322554] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.vendordata_dynamic_targets = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.322759] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.vendordata_jsonfile_path = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.322948] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api.vendordata_providers = ['StaticJSON'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.323159] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.backend = dogpile.cache.memcached {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.323329] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.backend_argument = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.323504] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.config_prefix = cache.oslo {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.323674] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.dead_timeout = 60.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.323839] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.debug_cache_backend = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.324015] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.enable_retry_client = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.324180] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.enable_socket_keepalive = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.324350] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.enabled = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.324513] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.enforce_fips_mode = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.324677] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.expiration_time = 600 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.324838] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.hashclient_retry_attempts = 2 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.325009] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.hashclient_retry_delay = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.325199] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.memcache_dead_retry = 300 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.325366] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.memcache_password = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.325532] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.325695] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.325858] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.memcache_pool_maxsize = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.326037] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.memcache_pool_unused_timeout = 60 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.326211] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.memcache_sasl_enabled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.326390] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.memcache_servers = ['localhost:11211'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.326554] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.memcache_socket_timeout = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.326712] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.memcache_username = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.326887] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.proxies = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.327047] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.redis_db = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.327209] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.redis_password = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.327406] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.redis_sentinel_service_name = mymaster {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.327592] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.327752] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.redis_server = localhost:6379 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.327917] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.redis_socket_timeout = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.328087] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.redis_username = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.328251] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.retry_attempts = 2 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.328417] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.retry_delay = 0.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.328578] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.socket_keepalive_count = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.328735] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.socket_keepalive_idle = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.328893] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.socket_keepalive_interval = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.329058] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.tls_allowed_ciphers = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.329219] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.tls_cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.329376] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.tls_certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.329535] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.tls_enabled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.329692] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cache.tls_keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.329859] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.auth_section = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.330049] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.auth_type = password {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.330218] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.330394] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.catalog_info = volumev3::publicURL {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.330553] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.330716] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.330876] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.cross_az_attach = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.331046] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.debug = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.331208] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.endpoint_template = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.331373] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.http_retries = 3 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.331534] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.331689] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.331858] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.os_region_name = RegionOne {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.332031] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.332195] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cinder.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.332367] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.332527] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.cpu_dedicated_set = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.332683] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.cpu_shared_set = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.332845] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.image_type_exclude_list = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.333014] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.live_migration_wait_for_vif_plug = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.333197] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.max_concurrent_disk_ops = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.333382] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.max_disk_devices_to_attach = -1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.333547] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.333715] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.333877] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.resource_provider_association_refresh = 300 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.334060] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.334235] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.shutdown_retry_interval = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.334417] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.334593] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] conductor.workers = 2 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.334772] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] console.allowed_origins = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.334933] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] console.ssl_ciphers = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.335113] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] console.ssl_minimum_version = default {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.335306] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] consoleauth.enforce_session_timeout = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.335480] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] consoleauth.token_ttl = 600 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.335645] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.335800] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.335962] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.336133] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.connect_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.336293] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.connect_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.336448] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.endpoint_override = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.336608] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.336764] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.336920] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.max_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.337084] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.min_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.337242] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.region_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.337398] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.retriable_status_codes = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.337553] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.service_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.337717] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.service_type = accelerator {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.337877] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.338040] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.status_code_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.338198] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.status_code_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.338357] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.338533] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.338691] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] cyborg.version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.338866] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.backend = sqlalchemy {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.339051] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.connection = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.339228] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.connection_debug = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.339420] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.connection_parameters = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.339591] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.connection_recycle_time = 3600 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.339756] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.connection_trace = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.339918] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.db_inc_retry_interval = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.340091] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.db_max_retries = 20 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.340255] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.db_max_retry_interval = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.340417] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.db_retry_interval = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.340576] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.max_overflow = 50 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.340736] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.max_pool_size = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.340895] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.max_retries = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.341072] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.mysql_sql_mode = TRADITIONAL {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.341232] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.mysql_wsrep_sync_wait = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.341390] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.pool_timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.341549] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.retry_interval = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.341703] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.slave_connection = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.341862] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.sqlite_synchronous = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.342029] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] database.use_db_reconnect = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.342208] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.backend = sqlalchemy {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.342376] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.connection = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.342537] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.connection_debug = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.342702] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.connection_parameters = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.342862] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.connection_recycle_time = 3600 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.343029] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.connection_trace = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.343207] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.db_inc_retry_interval = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.343383] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.db_max_retries = 20 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.343547] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.db_max_retry_interval = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.343711] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.db_retry_interval = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.343870] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.max_overflow = 50 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.344047] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.max_pool_size = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.344218] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.max_retries = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.344387] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.344546] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.mysql_wsrep_sync_wait = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.344703] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.pool_timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.344862] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.retry_interval = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.345023] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.slave_connection = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.345200] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] api_database.sqlite_synchronous = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.345393] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] devices.enabled_mdev_types = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.345573] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.345744] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ephemeral_storage_encryption.default_format = luks {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.345909] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ephemeral_storage_encryption.enabled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.346083] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ephemeral_storage_encryption.key_size = 512 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.346278] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.api_servers = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.346446] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.346608] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.346770] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.346926] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.connect_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.347094] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.connect_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.347257] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.debug = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.347423] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.default_trusted_certificate_ids = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.347584] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.enable_certificate_validation = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.347742] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.enable_rbd_download = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.347898] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.endpoint_override = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.348076] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.348241] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.349276] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.max_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.349276] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.min_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.349276] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.num_retries = 3 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.349276] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.rbd_ceph_conf = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.349276] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.rbd_connect_timeout = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.349468] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.rbd_pool = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.349692] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.rbd_user = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.349882] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.region_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.350076] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.retriable_status_codes = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.350335] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.service_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.350442] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.service_type = image {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.350656] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.350836] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.status_code_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.351091] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.status_code_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.351229] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.351427] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.351616] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.verify_glance_signatures = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.351796] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] glance.version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.351993] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] guestfs.debug = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.352210] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] mks.enabled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.352614] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.352811] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] image_cache.manager_interval = 2400 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.353045] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] image_cache.precache_concurrency = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.353224] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] image_cache.remove_unused_base_images = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.353427] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.353641] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.353831] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] image_cache.subdirectory_name = _base {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.354028] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.api_max_retries = 60 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.354209] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.api_retry_interval = 2 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.354376] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.auth_section = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.354541] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.auth_type = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.354734] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.354896] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.355070] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.355416] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.conductor_group = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.355512] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.connect_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.355782] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.connect_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.355908] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.endpoint_override = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.356175] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.356420] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.356631] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.max_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.356822] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.min_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.357078] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.peer_list = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.357164] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.region_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.357351] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.retriable_status_codes = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.357559] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.serial_console_state_timeout = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.357720] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.service_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.357950] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.service_type = baremetal {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.358160] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.shard = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.358428] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.358621] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.status_code_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.358849] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.status_code_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.358993] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.359260] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.359471] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ironic.version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.359699] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.359917] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] key_manager.fixed_key = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.360129] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.360258] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.barbican_api_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.360480] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.barbican_endpoint = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364017] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.barbican_endpoint_type = public {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364017] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.barbican_region_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364017] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364017] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364017] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364017] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364017] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364241] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.number_of_retries = 60 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364241] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.retry_delay = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364241] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.send_service_user_token = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364241] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364241] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364241] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.verify_ssl = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364241] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican.verify_ssl_path = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364464] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican_service_user.auth_section = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364464] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican_service_user.auth_type = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364464] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican_service_user.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364464] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican_service_user.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364464] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican_service_user.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364464] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican_service_user.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364464] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican_service_user.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364636] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican_service_user.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364636] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] barbican_service_user.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364636] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vault.approle_role_id = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364636] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vault.approle_secret_id = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364733] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vault.kv_mountpoint = secret {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.364957] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vault.kv_path = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.365023] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vault.kv_version = 2 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.365164] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vault.namespace = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.365345] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vault.root_token_id = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.365504] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vault.ssl_ca_crt_file = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.365671] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vault.timeout = 60.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.365829] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vault.use_ssl = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.365993] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.366221] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.auth_section = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.366401] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.auth_type = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.366558] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.366712] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.366870] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.367035] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.connect_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.367193] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.connect_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.367343] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.endpoint_override = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.367496] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.367647] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.367797] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.max_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.367946] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.min_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.368476] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.region_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.368476] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.retriable_status_codes = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.368476] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.service_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.368648] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.service_type = identity {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.368752] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.368930] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.status_code_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.369109] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.status_code_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.369273] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.369455] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.369613] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] keystone.version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.369810] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.connection_uri = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.369971] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.cpu_mode = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.370166] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.cpu_model_extra_flags = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.370342] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.cpu_models = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.370514] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.cpu_power_governor_high = performance {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.370680] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.cpu_power_governor_low = powersave {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.370841] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.cpu_power_management = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.371085] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.371189] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.device_detach_attempts = 8 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.371353] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.device_detach_timeout = 20 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.371516] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.disk_cachemodes = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.371753] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.disk_prefix = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.371987] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.enabled_perf_events = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.372239] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.file_backed_memory = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.372437] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.gid_maps = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.372600] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.hw_disk_discard = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.372815] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.hw_machine_type = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.373059] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.images_rbd_ceph_conf = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.373259] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.373429] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.373600] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.images_rbd_glance_store_name = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.373770] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.images_rbd_pool = rbd {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.373937] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.images_type = default {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.374111] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.images_volume_group = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.374276] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.inject_key = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.374448] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.inject_partition = -2 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.374610] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.inject_password = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.374771] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.iscsi_iface = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.374929] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.iser_use_multipath = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.375114] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_bandwidth = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.375313] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_completion_timeout = 800 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.375482] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_downtime = 500 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.375644] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_downtime_delay = 75 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.375804] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_downtime_steps = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.375961] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_inbound_addr = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.376135] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_permit_auto_converge = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.376320] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_permit_post_copy = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.376485] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_scheme = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.376660] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_timeout_action = abort {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.376823] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_tunnelled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.376983] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_uri = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.377158] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.live_migration_with_native_tls = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.377319] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.max_queues = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.377483] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.mem_stats_period_seconds = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.377723] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.377887] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.nfs_mount_options = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.378198] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.378375] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.num_aoe_discover_tries = 3 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.378540] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.num_iser_scan_tries = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.378702] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.num_memory_encrypted_guests = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.378864] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.num_nvme_discover_tries = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.379035] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.num_pcie_ports = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.379208] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.num_volume_scan_tries = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.379376] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.pmem_namespaces = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.379532] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.quobyte_client_cfg = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.379815] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.379987] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.rbd_connect_timeout = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.380183] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.380352] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.380512] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.rbd_secret_uuid = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.380669] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.rbd_user = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.380830] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.realtime_scheduler_priority = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.381011] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.remote_filesystem_transport = ssh {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.381176] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.rescue_image_id = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.381342] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.rescue_kernel_id = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.381497] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.rescue_ramdisk_id = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.381660] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.rng_dev_path = /dev/urandom {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.381816] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.rx_queue_size = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.381979] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.smbfs_mount_options = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.382267] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.382442] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.snapshot_compression = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.382601] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.snapshot_image_format = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.382814] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.382979] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.sparse_logical_volumes = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.383153] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.swtpm_enabled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.383322] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.swtpm_group = tss {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.383491] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.swtpm_user = tss {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.383657] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.sysinfo_serial = unique {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.383814] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.tb_cache_size = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.383971] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.tx_queue_size = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.384145] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.uid_maps = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.384314] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.use_virtio_for_bridges = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.384481] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.virt_type = kvm {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.384647] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.volume_clear = zero {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.384810] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.volume_clear_size = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.384975] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.volume_use_multipath = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.385160] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.vzstorage_cache_path = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.385335] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.385503] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.vzstorage_mount_group = qemu {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.385668] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.vzstorage_mount_opts = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.385833] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.386120] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.386299] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.vzstorage_mount_user = stack {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.386467] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.386641] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.auth_section = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.386813] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.auth_type = password {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.386973] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.387146] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.387312] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.387471] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.connect_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.387629] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.connect_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.387797] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.default_floating_pool = public {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.387955] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.endpoint_override = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.388128] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.extension_sync_interval = 600 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.388290] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.http_retries = 3 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.388449] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.388604] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.388763] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.max_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.388929] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.metadata_proxy_shared_secret = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.389098] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.min_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.389270] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.ovs_bridge = br-int {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.389437] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.physnets = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.389605] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.region_name = RegionOne {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.389765] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.retriable_status_codes = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.389933] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.service_metadata_proxy = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.390116] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.service_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.390293] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.service_type = network {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.390457] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.390615] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.status_code_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.390773] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.status_code_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.390932] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.391125] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.391290] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] neutron.version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.391464] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] notifications.bdms_in_notifications = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.391639] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] notifications.default_level = INFO {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.391812] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] notifications.notification_format = unversioned {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.391973] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] notifications.notify_on_state_change = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.392160] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.392335] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] pci.alias = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.392504] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] pci.device_spec = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.392668] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] pci.report_in_placement = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.392837] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.auth_section = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.393014] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.auth_type = password {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.393193] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.auth_url = http://10.180.1.21/identity {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.393351] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.393506] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.393667] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.393823] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.connect_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.393976] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.connect_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.394144] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.default_domain_id = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.394299] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.default_domain_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.394456] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.domain_id = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.394609] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.domain_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.394762] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.endpoint_override = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.394920] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.395093] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.395285] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.max_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.395443] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.min_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.395607] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.password = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.395764] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.project_domain_id = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.395927] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.project_domain_name = Default {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.396110] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.project_id = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.396284] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.project_name = service {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.396452] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.region_name = RegionOne {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.396612] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.retriable_status_codes = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.396769] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.service_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.396938] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.service_type = placement {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.397110] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.397271] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.status_code_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.397431] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.status_code_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.397591] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.system_scope = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.397747] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.397902] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.trust_id = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.398067] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.user_domain_id = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.398236] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.user_domain_name = Default {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.398396] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.user_id = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.398567] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.username = nova {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.398745] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.398904] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] placement.version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.399093] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.cores = 20 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.399262] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.count_usage_from_placement = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.399432] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.399604] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.injected_file_content_bytes = 10240 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.399770] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.injected_file_path_length = 255 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.399935] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.injected_files = 5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.400125] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.instances = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.400298] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.key_pairs = 100 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.400469] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.metadata_items = 128 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.400634] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.ram = 51200 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.400796] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.recheck_quota = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.400961] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.server_group_members = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.401141] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] quota.server_groups = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.401319] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.401485] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.401649] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] scheduler.image_metadata_prefilter = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.401812] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.401975] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] scheduler.max_attempts = 3 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.402157] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] scheduler.max_placement_results = 1000 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.402323] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.402488] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] scheduler.query_placement_for_image_type_support = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.402653] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.402826] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] scheduler.workers = 2 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.402994] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.403192] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.403374] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.403543] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.403708] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.403871] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.404041] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.404237] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.404407] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.host_subset_size = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.404570] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.404727] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.image_properties_default_architecture = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.404889] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.405071] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.isolated_hosts = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.405245] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.isolated_images = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.405408] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.max_instances_per_host = 50 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.405565] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.405724] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.405882] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.pci_in_placement = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.406051] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.406214] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.406376] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.406534] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.406691] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.406849] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.407016] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.track_instance_changes = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.407193] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.407362] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] metrics.required = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.407523] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] metrics.weight_multiplier = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.407686] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] metrics.weight_of_unavailable = -10000.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.407849] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] metrics.weight_setting = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.408172] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.408348] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] serial_console.enabled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.408524] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] serial_console.port_range = 10000:20000 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.408690] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.408857] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.409041] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] serial_console.serialproxy_port = 6083 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.409218] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] service_user.auth_section = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.409393] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] service_user.auth_type = password {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.409553] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] service_user.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.409707] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] service_user.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.409870] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] service_user.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.410039] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] service_user.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.410201] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] service_user.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.410372] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] service_user.send_service_user_token = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.410535] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] service_user.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.410704] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] service_user.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.410874] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.agent_enabled = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.411048] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.enabled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.411364] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.411554] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.html5proxy_host = 0.0.0.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.411721] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.html5proxy_port = 6082 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.411882] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.image_compression = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.412049] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.jpeg_compression = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.412211] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.playback_compression = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.412373] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.require_secure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.412539] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.server_listen = 127.0.0.1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.412705] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.412860] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.streaming_mode = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.413032] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] spice.zlib_compression = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.413212] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] upgrade_levels.baseapi = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.413384] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] upgrade_levels.compute = auto {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.413539] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] upgrade_levels.conductor = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.413694] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] upgrade_levels.scheduler = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.413858] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vendordata_dynamic_auth.auth_section = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.414028] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vendordata_dynamic_auth.auth_type = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.414189] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vendordata_dynamic_auth.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.414348] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vendordata_dynamic_auth.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.414509] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vendordata_dynamic_auth.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.414668] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vendordata_dynamic_auth.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.414823] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vendordata_dynamic_auth.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.414983] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vendordata_dynamic_auth.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.415150] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vendordata_dynamic_auth.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.415335] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.api_retry_count = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.415508] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.ca_file = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.415680] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.cache_prefix = devstack-image-cache {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.415846] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.cluster_name = testcl1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.416015] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.connection_pool_size = 10 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.416181] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.console_delay_seconds = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.416349] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.datastore_regex = ^datastore.* {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.416547] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.416717] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.host_password = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.416883] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.host_port = 443 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.417070] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.host_username = administrator@vsphere.local {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.417248] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.insecure = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.417411] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.integration_bridge = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.417576] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.maximum_objects = 100 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.417734] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.pbm_default_policy = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.417895] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.pbm_enabled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.418063] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.pbm_wsdl_location = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.418233] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.418392] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.serial_port_proxy_uri = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.418555] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.serial_port_service_uri = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.418721] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.task_poll_interval = 0.5 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.418891] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.use_linked_clone = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.419068] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.vnc_keymap = en-us {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.419235] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.vnc_port = 5900 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.419403] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vmware.vnc_port_total = 10000 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.419585] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vnc.auth_schemes = ['none'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.419756] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vnc.enabled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.420054] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.420240] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.420412] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vnc.novncproxy_port = 6080 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.420588] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vnc.server_listen = 127.0.0.1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.420761] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.420922] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vnc.vencrypt_ca_certs = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.421091] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vnc.vencrypt_client_cert = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.421252] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vnc.vencrypt_client_key = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.421430] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.421595] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.disable_deep_image_inspection = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.421760] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.disable_fallback_pcpu_query = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.421921] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.disable_group_policy_check_upcall = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.422103] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.422274] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.disable_rootwrap = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.422441] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.enable_numa_live_migration = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.422599] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.422759] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.422921] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.handle_virt_lifecycle_events = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.423093] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.libvirt_disable_apic = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.423259] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.never_download_image_if_on_rbd = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.423425] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.423588] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.423748] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.423910] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.424084] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.424249] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.424413] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.424572] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.424735] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.424918] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.425099] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] wsgi.client_socket_timeout = 900 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.425274] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] wsgi.default_pool_size = 1000 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.425443] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] wsgi.keep_alive = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.425608] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] wsgi.max_header_line = 16384 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.425767] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] wsgi.secure_proxy_ssl_header = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.425928] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] wsgi.ssl_ca_file = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.426095] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] wsgi.ssl_cert_file = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.426261] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] wsgi.ssl_key_file = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.426447] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] wsgi.tcp_keepidle = 600 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.426630] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.426796] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] zvm.ca_file = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.426955] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] zvm.cloud_connector_url = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.427259] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.427441] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] zvm.reachable_timeout = 300 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.427623] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_policy.enforce_new_defaults = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.428024] env[63293]: WARNING oslo_config.cfg [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 470.428204] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_policy.enforce_scope = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.428380] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_policy.policy_default_rule = default {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.428564] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.428737] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_policy.policy_file = policy.yaml {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.428913] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.429085] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.429249] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.429409] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.429570] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.429738] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.429912] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.430098] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler.connection_string = messaging:// {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.430269] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler.enabled = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.430440] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler.es_doc_type = notification {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.430604] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler.es_scroll_size = 10000 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.430769] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler.es_scroll_time = 2m {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.430929] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler.filter_error_trace = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.431108] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler.hmac_keys = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.431277] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler.sentinel_service_name = mymaster {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.431445] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler.socket_timeout = 0.1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.431605] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler.trace_requests = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.431763] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler.trace_sqlalchemy = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.431937] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler_jaeger.process_tags = {} {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.432118] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler_jaeger.service_name_prefix = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.432288] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] profiler_otlp.service_name_prefix = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.432454] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] remote_debug.host = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.432611] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] remote_debug.port = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.432790] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.432952] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.433128] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.433293] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.433455] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.433612] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.433770] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.433928] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.434097] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.434270] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.hostname = devstack {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.434429] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.434597] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.434760] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.434927] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.435104] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.435273] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.435437] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.435606] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.435768] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.435927] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.436103] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.436272] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.436434] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.436598] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.436759] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.436919] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.437100] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.437267] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.437436] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.437602] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.ssl = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.437770] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.437936] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.438108] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.438278] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.438446] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.ssl_version = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.438604] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.438788] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.438953] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_notifications.retry = -1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.439144] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.439321] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_messaging_notifications.transport_url = **** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.439491] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.auth_section = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.439650] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.auth_type = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.439805] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.cafile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.439959] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.certfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.440130] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.collect_timing = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.440290] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.connect_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.440447] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.connect_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.440602] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.endpoint_id = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.440753] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.endpoint_override = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.440910] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.insecure = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.441072] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.keyfile = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.441229] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.max_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.441383] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.min_version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.441536] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.region_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.441693] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.retriable_status_codes = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.441846] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.service_name = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.441999] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.service_type = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.442185] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.split_loggers = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.442345] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.status_code_retries = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.442500] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.status_code_retry_delay = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.442654] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.timeout = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.442808] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.valid_interfaces = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.442960] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_limit.version = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.443135] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_reports.file_event_handler = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.443300] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_reports.file_event_handler_interval = 1 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.443455] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] oslo_reports.log_dir = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.443621] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.443777] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_linux_bridge_privileged.group = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.443933] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.444112] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.444278] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.444437] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_linux_bridge_privileged.user = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.444603] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.444760] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_ovs_privileged.group = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.444915] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_ovs_privileged.helper_command = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.445086] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.445253] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.445409] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] vif_plug_ovs_privileged.user = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.445576] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_linux_bridge.flat_interface = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.445755] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.445926] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.446108] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.446283] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.446450] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.446617] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.446780] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_linux_bridge.vlan_interface = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.446957] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.447152] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_ovs.isolate_vif = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.447323] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.447493] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.447662] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.447830] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_ovs.ovsdb_interface = native {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.447992] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] os_vif_ovs.per_port_bridge = False {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.448178] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] privsep_osbrick.capabilities = [21] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.448340] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] privsep_osbrick.group = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.448498] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] privsep_osbrick.helper_command = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.448660] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.448823] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] privsep_osbrick.thread_pool_size = 8 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.448980] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] privsep_osbrick.user = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.449163] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.449326] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] nova_sys_admin.group = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.449482] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] nova_sys_admin.helper_command = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.449642] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.449799] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] nova_sys_admin.thread_pool_size = 8 {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.449953] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] nova_sys_admin.user = None {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.450092] env[63293]: DEBUG oslo_service.service [None req-b0f9610d-dbe1-4ae1-a494-3fedde0b095e None None] ******************************************************************************** {{(pid=63293) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 470.450565] env[63293]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 470.954135] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Getting list of instances from cluster (obj){ [ 470.954135] env[63293]: value = "domain-c8" [ 470.954135] env[63293]: _type = "ClusterComputeResource" [ 470.954135] env[63293]: } {{(pid=63293) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 470.955380] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4b9f8a-3574-44ec-a5f5-23bdb0e60658 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.964169] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Got total of 0 instances {{(pid=63293) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 470.964688] env[63293]: WARNING nova.virt.vmwareapi.driver [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 470.965179] env[63293]: INFO nova.virt.node [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Generated node identity 619725c3-6194-4724-94b1-1dd499be55d0 [ 470.965418] env[63293]: INFO nova.virt.node [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Wrote node identity 619725c3-6194-4724-94b1-1dd499be55d0 to /opt/stack/data/n-cpu-1/compute_id [ 471.468600] env[63293]: WARNING nova.compute.manager [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Compute nodes ['619725c3-6194-4724-94b1-1dd499be55d0'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 472.474522] env[63293]: INFO nova.compute.manager [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 473.479935] env[63293]: WARNING nova.compute.manager [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 473.480305] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 473.480429] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 473.480577] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 473.480729] env[63293]: DEBUG nova.compute.resource_tracker [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63293) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 473.481751] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55d6f6e-9c65-4340-9dfd-f9f4809f423e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.490263] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58124ae0-39f3-44aa-85a0-0b9cc953a2b6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.503844] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8ed241-de5d-4b16-967f-50df795b8366 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.510632] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbe9994-56b5-49a6-b57e-3953fbf186f6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.540068] env[63293]: DEBUG nova.compute.resource_tracker [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181521MB free_disk=132GB free_vcpus=48 pci_devices=None {{(pid=63293) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 473.540253] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 473.540406] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 474.043729] env[63293]: WARNING nova.compute.resource_tracker [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] No compute node record for cpu-1:619725c3-6194-4724-94b1-1dd499be55d0: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 619725c3-6194-4724-94b1-1dd499be55d0 could not be found. [ 474.546941] env[63293]: INFO nova.compute.resource_tracker [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 619725c3-6194-4724-94b1-1dd499be55d0 [ 476.054729] env[63293]: DEBUG nova.compute.resource_tracker [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 476.055152] env[63293]: DEBUG nova.compute.resource_tracker [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 476.205884] env[63293]: INFO nova.scheduler.client.report [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] [req-ab504226-33d6-47df-a0f3-d3b6e611eea4] Created resource provider record via placement API for resource provider with UUID 619725c3-6194-4724-94b1-1dd499be55d0 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 476.221107] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e936ffc-5d14-4395-94e9-84637d7a7914 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.228900] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8583a360-7ad8-49a9-95ba-21cd19089587 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.258801] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd384b98-c25c-4351-8549-665fa9c861db {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.265973] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d644f3-2ce3-4e5f-bbb8-bcf94d940c04 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.278678] env[63293]: DEBUG nova.compute.provider_tree [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 476.825069] env[63293]: DEBUG nova.scheduler.client.report [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Updated inventory for provider 619725c3-6194-4724-94b1-1dd499be55d0 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 476.825341] env[63293]: DEBUG nova.compute.provider_tree [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Updating resource provider 619725c3-6194-4724-94b1-1dd499be55d0 generation from 0 to 1 during operation: update_inventory {{(pid=63293) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 476.825472] env[63293]: DEBUG nova.compute.provider_tree [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 476.877897] env[63293]: DEBUG nova.compute.provider_tree [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Updating resource provider 619725c3-6194-4724-94b1-1dd499be55d0 generation from 1 to 2 during operation: update_traits {{(pid=63293) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 477.382937] env[63293]: DEBUG nova.compute.resource_tracker [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 477.383391] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.843s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 477.383391] env[63293]: DEBUG nova.service [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Creating RPC server for service compute {{(pid=63293) start /opt/stack/nova/nova/service.py:186}} [ 477.395735] env[63293]: DEBUG nova.service [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] Join ServiceGroup membership for this service compute {{(pid=63293) start /opt/stack/nova/nova/service.py:203}} [ 477.395926] env[63293]: DEBUG nova.servicegroup.drivers.db [None req-cf2081e4-9f9a-4781-b0d7-c1833579069a None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=63293) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 481.397769] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._sync_power_states {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 481.901423] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Getting list of instances from cluster (obj){ [ 481.901423] env[63293]: value = "domain-c8" [ 481.901423] env[63293]: _type = "ClusterComputeResource" [ 481.901423] env[63293]: } {{(pid=63293) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 481.902607] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2ea2e9-0422-429d-92b4-3246d9cd03a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.911012] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Got total of 0 instances {{(pid=63293) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 481.911259] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 481.911565] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Getting list of instances from cluster (obj){ [ 481.911565] env[63293]: value = "domain-c8" [ 481.911565] env[63293]: _type = "ClusterComputeResource" [ 481.911565] env[63293]: } {{(pid=63293) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 481.912391] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa2e57d-6198-4518-9e45-672d75637b2c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.919541] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Got total of 0 instances {{(pid=63293) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 515.514192] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquiring lock "d9d24a3a-078a-49a6-a02c-3fb0e09ef772" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.514513] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "d9d24a3a-078a-49a6-a02c-3fb0e09ef772" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.016960] env[63293]: DEBUG nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.576511] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.576756] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.580671] env[63293]: INFO nova.compute.claims [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 517.094684] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquiring lock "e80209b3-273c-4045-ad8a-2a3c4c48200a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.094957] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "e80209b3-273c-4045-ad8a-2a3c4c48200a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.179778] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Acquiring lock "f9e64134-f233-48a3-afc3-be7765a67c35" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.179778] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Lock "f9e64134-f233-48a3-afc3-be7765a67c35" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.601173] env[63293]: DEBUG nova.compute.manager [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.677422] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa84e92-8f12-41a0-8261-5ae1f240700f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.682931] env[63293]: DEBUG nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.691901] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3711ad92-d386-4b22-b884-5cce775576a3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.731441] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00260b85-d8ea-403f-824e-4f4c8b74eef4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.741366] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb40af1-bcb6-4ff6-97c0-ff8434553252 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.756736] env[63293]: DEBUG nova.compute.provider_tree [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 518.131574] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.217330] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.260224] env[63293]: DEBUG nova.scheduler.client.report [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 518.436846] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Acquiring lock "c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.438109] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Lock "c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.768061] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.191s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.769846] env[63293]: DEBUG nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 518.774889] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.645s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.777019] env[63293]: INFO nova.compute.claims [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.942584] env[63293]: DEBUG nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 519.286563] env[63293]: DEBUG nova.compute.utils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 519.287481] env[63293]: DEBUG nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 519.287878] env[63293]: DEBUG nova.network.neutron [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 519.484863] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.790566] env[63293]: DEBUG nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 519.848450] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Acquiring lock "aa4611e2-6254-4983-8565-899aea616726" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.848450] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Lock "aa4611e2-6254-4983-8565-899aea616726" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.968356] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2983f4-38d0-4a4d-abd9-80d39fd49300 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.976599] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a96b8c-ce0d-4a97-915f-fb27a8809ad9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.021868] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852aa3c9-3b44-42ae-af7d-4bd129e26356 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.029791] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af1ee77-c190-49f3-812e-8f883576b1d8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.049137] env[63293]: DEBUG nova.compute.provider_tree [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 520.138616] env[63293]: DEBUG nova.policy [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a31e6add8ee54ce69bc3ffd9a1056f1a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'db20579b85394de3bfb7afdd7dd78207', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 520.350980] env[63293]: DEBUG nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.554425] env[63293]: DEBUG nova.scheduler.client.report [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 520.570129] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquiring lock "c9221012-9f13-4881-9a0a-1b9c1c84d910" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.571398] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "c9221012-9f13-4881-9a0a-1b9c1c84d910" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.806295] env[63293]: DEBUG nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 520.846356] env[63293]: DEBUG nova.virt.hardware [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 520.847016] env[63293]: DEBUG nova.virt.hardware [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 520.847016] env[63293]: DEBUG nova.virt.hardware [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 520.847016] env[63293]: DEBUG nova.virt.hardware [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 520.850239] env[63293]: DEBUG nova.virt.hardware [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 520.850239] env[63293]: DEBUG nova.virt.hardware [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 520.850239] env[63293]: DEBUG nova.virt.hardware [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 520.850239] env[63293]: DEBUG nova.virt.hardware [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 520.850565] env[63293]: DEBUG nova.virt.hardware [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 520.850738] env[63293]: DEBUG nova.virt.hardware [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 520.850907] env[63293]: DEBUG nova.virt.hardware [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 520.852247] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3950a71-6109-4835-9e5d-f72278f58c42 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.866849] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630a0e40-46b7-4157-aa7a-b7e48737627e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.890297] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e43ae8-771f-4894-82ba-ba7f720d5e25 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.904291] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.986849] env[63293]: DEBUG nova.network.neutron [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Successfully created port: 6a6d81e5-3c63-425b-938b-2294443c2fd4 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 521.063471] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.288s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.063738] env[63293]: DEBUG nova.compute.manager [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 521.069816] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.854s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.074102] env[63293]: INFO nova.compute.claims [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 521.076283] env[63293]: DEBUG nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 521.577773] env[63293]: DEBUG nova.compute.utils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.582161] env[63293]: DEBUG nova.compute.manager [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Not allocating networking since 'none' was specified. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 521.611138] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.084322] env[63293]: DEBUG nova.compute.manager [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 522.185546] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "f255ca12-7110-4f73-a6ff-d6bcc122486b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.185546] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "f255ca12-7110-4f73-a6ff-d6bcc122486b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.244660] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e246e05-9811-4e9b-87c8-efb48d746f52 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.255303] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5c2036-c0bc-477e-85aa-370c7440ad12 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.291853] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6442b335-6384-4c80-973e-c8e99070665d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.299847] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f2878f-3f3d-4aa6-bb0a-dc6e53b8b591 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.314892] env[63293]: DEBUG nova.compute.provider_tree [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.690166] env[63293]: DEBUG nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 522.818618] env[63293]: DEBUG nova.scheduler.client.report [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 522.942171] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Acquiring lock "dddc6041-4436-48da-8ed8-45be063b566b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.942543] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Lock "dddc6041-4436-48da-8ed8-45be063b566b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.098759] env[63293]: DEBUG nova.compute.manager [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 523.126467] env[63293]: DEBUG nova.virt.hardware [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 523.126731] env[63293]: DEBUG nova.virt.hardware [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 523.126912] env[63293]: DEBUG nova.virt.hardware [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 523.127104] env[63293]: DEBUG nova.virt.hardware [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 523.127597] env[63293]: DEBUG nova.virt.hardware [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 523.127763] env[63293]: DEBUG nova.virt.hardware [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 523.127976] env[63293]: DEBUG nova.virt.hardware [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 523.128237] env[63293]: DEBUG nova.virt.hardware [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 523.128315] env[63293]: DEBUG nova.virt.hardware [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 523.128477] env[63293]: DEBUG nova.virt.hardware [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 523.128636] env[63293]: DEBUG nova.virt.hardware [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 523.130359] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c96540-d089-4888-a3f1-2a4d87a60c71 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.138903] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087a6ba8-204b-4ca7-bc65-174883d4b81f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.153753] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 523.165012] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 523.169606] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-444cbdb8-eb19-4359-a93d-099d5d0df045 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.180240] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Created folder: OpenStack in parent group-v4. [ 523.180437] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Creating folder: Project (5fb465463d2e40b79e78c67024ff71bb). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 523.180676] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-02e6c4a9-92b2-411a-a415-36d50c469399 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.192216] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Created folder: Project (5fb465463d2e40b79e78c67024ff71bb) in parent group-v283678. [ 523.192545] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Creating folder: Instances. Parent ref: group-v283679. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 523.192644] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1e169aa4-7fa6-42f0-84b4-9a5fad2f388b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.207941] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Created folder: Instances in parent group-v283679. [ 523.208367] env[63293]: DEBUG oslo.service.loopingcall [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 523.208611] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 523.208837] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa522fe1-328a-41da-a80d-93a92a58a9df {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.227061] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.232880] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 523.232880] env[63293]: value = "task-1327407" [ 523.232880] env[63293]: _type = "Task" [ 523.232880] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.242766] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327407, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.330730] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.335293] env[63293]: DEBUG nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 523.337236] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.853s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.339623] env[63293]: INFO nova.compute.claims [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 523.445322] env[63293]: DEBUG nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 523.634937] env[63293]: ERROR nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6a6d81e5-3c63-425b-938b-2294443c2fd4, please check neutron logs for more information. [ 523.634937] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.634937] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.634937] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.634937] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.634937] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.634937] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.634937] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.634937] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.634937] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 523.634937] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.634937] env[63293]: ERROR nova.compute.manager raise self.value [ 523.634937] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.634937] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.634937] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.634937] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.635746] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.635746] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.635746] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6a6d81e5-3c63-425b-938b-2294443c2fd4, please check neutron logs for more information. [ 523.635746] env[63293]: ERROR nova.compute.manager [ 523.635746] env[63293]: Traceback (most recent call last): [ 523.635746] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.635746] env[63293]: listener.cb(fileno) [ 523.635746] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.635746] env[63293]: result = function(*args, **kwargs) [ 523.635746] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.635746] env[63293]: return func(*args, **kwargs) [ 523.635746] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.635746] env[63293]: raise e [ 523.635746] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.635746] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 523.635746] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.635746] env[63293]: created_port_ids = self._update_ports_for_instance( [ 523.635746] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.635746] env[63293]: with excutils.save_and_reraise_exception(): [ 523.635746] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.635746] env[63293]: self.force_reraise() [ 523.635746] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.635746] env[63293]: raise self.value [ 523.635746] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.635746] env[63293]: updated_port = self._update_port( [ 523.635746] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.635746] env[63293]: _ensure_no_port_binding_failure(port) [ 523.635746] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.635746] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.636584] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 6a6d81e5-3c63-425b-938b-2294443c2fd4, please check neutron logs for more information. [ 523.636584] env[63293]: Removing descriptor: 15 [ 523.639434] env[63293]: ERROR nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6a6d81e5-3c63-425b-938b-2294443c2fd4, please check neutron logs for more information. [ 523.639434] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Traceback (most recent call last): [ 523.639434] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.639434] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] yield resources [ 523.639434] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.639434] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] self.driver.spawn(context, instance, image_meta, [ 523.639434] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 523.639434] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.639434] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.639434] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] vm_ref = self.build_virtual_machine(instance, [ 523.639434] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] for vif in network_info: [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] return self._sync_wrapper(fn, *args, **kwargs) [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] self.wait() [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] self[:] = self._gt.wait() [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] return self._exit_event.wait() [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.640724] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] result = hub.switch() [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] return self.greenlet.switch() [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] result = function(*args, **kwargs) [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] return func(*args, **kwargs) [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] raise e [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] nwinfo = self.network_api.allocate_for_instance( [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] created_port_ids = self._update_ports_for_instance( [ 523.641224] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] with excutils.save_and_reraise_exception(): [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] self.force_reraise() [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] raise self.value [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] updated_port = self._update_port( [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] _ensure_no_port_binding_failure(port) [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] raise exception.PortBindingFailed(port_id=port['id']) [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] nova.exception.PortBindingFailed: Binding failed for port 6a6d81e5-3c63-425b-938b-2294443c2fd4, please check neutron logs for more information. [ 523.641598] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] [ 523.641949] env[63293]: INFO nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Terminating instance [ 523.642912] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquiring lock "refresh_cache-d9d24a3a-078a-49a6-a02c-3fb0e09ef772" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.642912] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquired lock "refresh_cache-d9d24a3a-078a-49a6-a02c-3fb0e09ef772" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.644087] env[63293]: DEBUG nova.network.neutron [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 523.744626] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327407, 'name': CreateVM_Task, 'duration_secs': 0.341277} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.744626] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 523.746134] env[63293]: DEBUG oslo_vmware.service [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375f8d31-bf14-40da-a851-8303bdeda7f1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.753451] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.754454] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.755532] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 523.755828] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e7a39c1-d664-476f-b9f7-5d5ac38ed875 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.760801] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 523.760801] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52828b4c-7959-6a92-cd8a-3f73f32b8228" [ 523.760801] env[63293]: _type = "Task" [ 523.760801] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.770426] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52828b4c-7959-6a92-cd8a-3f73f32b8228, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.850926] env[63293]: DEBUG nova.compute.utils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.855990] env[63293]: DEBUG nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 523.856368] env[63293]: DEBUG nova.network.neutron [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 523.907432] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquiring lock "9982e59e-65af-44c7-bdd9-056565107a76" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.907670] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "9982e59e-65af-44c7-bdd9-056565107a76" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.977701] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.124079] env[63293]: DEBUG nova.policy [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '157fb31c4e03475eb4dcb541e70baed0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a229ad83126f47baa1f92fa25fd7bcc6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 524.177879] env[63293]: DEBUG nova.network.neutron [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.278413] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.278951] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 524.279390] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.279874] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.280557] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 524.281706] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f01ba31-b04a-4a55-bd2c-3451d69ff578 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.302723] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 524.302723] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 524.303596] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308bd850-b4cf-4c4f-89f7-e66d075e4847 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.312722] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8773caab-5e43-44ca-8062-e5571b375677 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.320339] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 524.320339] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52359df8-77bd-4433-74ad-a1d05721b974" [ 524.320339] env[63293]: _type = "Task" [ 524.320339] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.330915] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52359df8-77bd-4433-74ad-a1d05721b974, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.357678] env[63293]: DEBUG nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 524.385427] env[63293]: DEBUG nova.network.neutron [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.410284] env[63293]: DEBUG nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.583050] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d249e3e0-e6f5-4c4b-a0d4-0c5ca2f344e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.592864] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2dccee-33b0-4d3a-9def-bc087900998a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.627479] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8dcb88a-dcf0-4f9d-855f-75548ad125d8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.634990] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d44cae5-d681-4ffe-9c49-34977b1922ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.650734] env[63293]: DEBUG nova.compute.provider_tree [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.735598] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "dccd4149-a888-4910-95c8-8d0028338dee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.736939] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "dccd4149-a888-4910-95c8-8d0028338dee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.833282] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Preparing fetch location {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 524.833282] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Creating directory with path [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 524.833282] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a41817a3-79e2-4384-85b3-47f83cbc0694 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.854378] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Created directory with path [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 524.854588] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Fetch image to [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 524.855116] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Downloading image file data 1427aa66-93c7-49c6-9e28-dc7fe851dced to [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk on the data store datastore2 {{(pid=63293) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 524.856074] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2412b20b-a90c-459e-9d9d-53c62281d16e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.864272] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656865d4-2872-434d-8de4-0eb3d76e92a0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.876416] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828e7d39-73ca-49a2-9c5b-bce91e46d8c8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.910833] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Releasing lock "refresh_cache-d9d24a3a-078a-49a6-a02c-3fb0e09ef772" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.911260] env[63293]: DEBUG nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 524.911445] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 524.912319] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-97083e54-0871-407a-9d38-b7ed3ade6fe9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.914895] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de54a8d-afcc-4799-afd4-f30fd753e756 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.930038] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f1d0c7-4991-428d-9b06-ff6d10d82a8d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.944922] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.949327] env[63293]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-d8129d4a-efb6-4035-b01e-44338328ddd2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.956976] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d9d24a3a-078a-49a6-a02c-3fb0e09ef772 could not be found. [ 524.957228] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 524.957615] env[63293]: INFO nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Took 0.05 seconds to destroy the instance on the hypervisor. [ 524.957884] env[63293]: DEBUG oslo.service.loopingcall [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 524.958150] env[63293]: DEBUG nova.compute.manager [-] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.958250] env[63293]: DEBUG nova.network.neutron [-] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 524.977841] env[63293]: DEBUG nova.network.neutron [-] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.981727] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Downloading image file data 1427aa66-93c7-49c6-9e28-dc7fe851dced to the data store datastore2 {{(pid=63293) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 525.064442] env[63293]: DEBUG oslo_vmware.rw_handles [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63293) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 525.125795] env[63293]: DEBUG nova.network.neutron [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Successfully created port: ead43854-09d6-4e20-ab13-c5e48dd12379 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 525.153019] env[63293]: DEBUG nova.scheduler.client.report [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 525.241033] env[63293]: DEBUG nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 525.376799] env[63293]: DEBUG nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 525.394988] env[63293]: DEBUG nova.compute.manager [req-66193d27-481f-4e98-aa6a-97ac30df7e8d req-5efb04e5-9226-491f-ac22-196831d12783 service nova] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Received event network-changed-6a6d81e5-3c63-425b-938b-2294443c2fd4 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.398270] env[63293]: DEBUG nova.compute.manager [req-66193d27-481f-4e98-aa6a-97ac30df7e8d req-5efb04e5-9226-491f-ac22-196831d12783 service nova] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Refreshing instance network info cache due to event network-changed-6a6d81e5-3c63-425b-938b-2294443c2fd4. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 525.398270] env[63293]: DEBUG oslo_concurrency.lockutils [req-66193d27-481f-4e98-aa6a-97ac30df7e8d req-5efb04e5-9226-491f-ac22-196831d12783 service nova] Acquiring lock "refresh_cache-d9d24a3a-078a-49a6-a02c-3fb0e09ef772" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.398270] env[63293]: DEBUG oslo_concurrency.lockutils [req-66193d27-481f-4e98-aa6a-97ac30df7e8d req-5efb04e5-9226-491f-ac22-196831d12783 service nova] Acquired lock "refresh_cache-d9d24a3a-078a-49a6-a02c-3fb0e09ef772" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.398270] env[63293]: DEBUG nova.network.neutron [req-66193d27-481f-4e98-aa6a-97ac30df7e8d req-5efb04e5-9226-491f-ac22-196831d12783 service nova] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Refreshing network info cache for port 6a6d81e5-3c63-425b-938b-2294443c2fd4 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 525.421340] env[63293]: DEBUG nova.virt.hardware [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 525.422307] env[63293]: DEBUG nova.virt.hardware [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 525.422307] env[63293]: DEBUG nova.virt.hardware [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 525.422307] env[63293]: DEBUG nova.virt.hardware [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 525.422492] env[63293]: DEBUG nova.virt.hardware [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 525.422625] env[63293]: DEBUG nova.virt.hardware [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 525.422875] env[63293]: DEBUG nova.virt.hardware [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 525.423050] env[63293]: DEBUG nova.virt.hardware [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 525.423620] env[63293]: DEBUG nova.virt.hardware [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 525.423620] env[63293]: DEBUG nova.virt.hardware [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 525.423620] env[63293]: DEBUG nova.virt.hardware [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 525.425552] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea95fcf3-1aa1-439f-a8db-cf857e9add67 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.446218] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d658227-8ae3-4d97-b32f-b8b3b31facc2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.485139] env[63293]: DEBUG nova.network.neutron [-] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.665377] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.328s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.665892] env[63293]: DEBUG nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 525.670619] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.766s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.672144] env[63293]: INFO nova.compute.claims [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 525.786719] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "94ad1eb8-d583-45a1-a94e-e13eb57b2b87" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.786970] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "94ad1eb8-d583-45a1-a94e-e13eb57b2b87" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.788260] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.798514] env[63293]: DEBUG oslo_vmware.rw_handles [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Completed reading data from the image iterator. {{(pid=63293) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 525.798514] env[63293]: DEBUG oslo_vmware.rw_handles [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 525.822921] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "43582f6d-75fb-4734-9fb0-b71db3ddf74c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.822921] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "43582f6d-75fb-4734-9fb0-b71db3ddf74c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.854243] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Downloaded image file data 1427aa66-93c7-49c6-9e28-dc7fe851dced to vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk on the data store datastore2 {{(pid=63293) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 525.856215] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Caching image {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 525.856555] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Copying Virtual Disk [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk to [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 525.860486] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-626dd742-f9a1-4047-af4c-8be74912ab9b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.872175] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 525.872175] env[63293]: value = "task-1327408" [ 525.872175] env[63293]: _type = "Task" [ 525.872175] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.881495] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.881811] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327408, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.882024] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.882210] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Starting heal instance info cache {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 525.882331] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Rebuilding the list of instances to heal {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 525.887686] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "0b381f4c-372f-41c8-9661-7e6bab6dd306" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.888237] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "0b381f4c-372f-41c8-9661-7e6bab6dd306" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.931250] env[63293]: DEBUG nova.network.neutron [req-66193d27-481f-4e98-aa6a-97ac30df7e8d req-5efb04e5-9226-491f-ac22-196831d12783 service nova] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.989358] env[63293]: INFO nova.compute.manager [-] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Took 1.03 seconds to deallocate network for instance. [ 526.000916] env[63293]: DEBUG nova.compute.claims [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 526.005032] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.005032] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "4dce2ec6-6f7b-4294-a32e-4cd222ae3492" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.006291] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "4dce2ec6-6f7b-4294-a32e-4cd222ae3492" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.010433] env[63293]: DEBUG nova.network.neutron [req-66193d27-481f-4e98-aa6a-97ac30df7e8d req-5efb04e5-9226-491f-ac22-196831d12783 service nova] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.174231] env[63293]: DEBUG nova.compute.utils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 526.178060] env[63293]: DEBUG nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 526.178343] env[63293]: DEBUG nova.network.neutron [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 526.243926] env[63293]: DEBUG nova.policy [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3024a0ab34842669be4e712e72dfac8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '67fb2ed2cc924227be0a024ebaf35d60', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 526.388265] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327408, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.392203] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 526.392203] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 526.392203] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 526.392203] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 526.392203] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: aa4611e2-6254-4983-8565-899aea616726] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 526.392203] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Didn't find any instances for network info cache update. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 526.392472] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.393249] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.393810] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.395069] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.395069] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.395069] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.395069] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63293) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 526.395295] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.402946] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquiring lock "1ac2f52a-27ef-4ce4-9540-288658f08344" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.403228] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "1ac2f52a-27ef-4ce4-9540-288658f08344" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.512866] env[63293]: DEBUG oslo_concurrency.lockutils [req-66193d27-481f-4e98-aa6a-97ac30df7e8d req-5efb04e5-9226-491f-ac22-196831d12783 service nova] Releasing lock "refresh_cache-d9d24a3a-078a-49a6-a02c-3fb0e09ef772" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.682261] env[63293]: DEBUG nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 526.881736] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327408, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.704602} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.881997] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Copied Virtual Disk [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk to [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 526.882418] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleting the datastore file [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 526.882622] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a61769d1-4466-4131-88be-7b06cff62c9f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.893832] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 526.893832] env[63293]: value = "task-1327409" [ 526.893832] env[63293]: _type = "Task" [ 526.893832] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.906803] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.906803] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327409, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.943633] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb06d95-6453-455d-8bd7-3e2df763eb85 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.952574] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e22b8d-834b-45e0-807e-f1d1628cb735 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.982221] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd3ff51-adab-4482-88c9-a81fa8e9a74e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.989850] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610eb1f3-f81d-424e-9b46-c0c8fcf8bf90 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.004332] env[63293]: DEBUG nova.compute.provider_tree [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.323142] env[63293]: DEBUG nova.network.neutron [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Successfully created port: a9f614a5-9c36-44bc-8b23-6b8cc5e475d9 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 527.405353] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327409, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031409} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.405671] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 527.405863] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Moving file from [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b/1427aa66-93c7-49c6-9e28-dc7fe851dced to [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced. {{(pid=63293) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 527.406094] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-26ae730c-1d2c-47f4-956a-59356e645c40 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.412778] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 527.412778] env[63293]: value = "task-1327410" [ 527.412778] env[63293]: _type = "Task" [ 527.412778] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.421777] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327410, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.424629] env[63293]: ERROR nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ead43854-09d6-4e20-ab13-c5e48dd12379, please check neutron logs for more information. [ 527.424629] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 527.424629] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.424629] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 527.424629] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.424629] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 527.424629] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.424629] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 527.424629] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.424629] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 527.424629] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.424629] env[63293]: ERROR nova.compute.manager raise self.value [ 527.424629] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.424629] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 527.424629] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.424629] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 527.425163] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.425163] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 527.425163] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ead43854-09d6-4e20-ab13-c5e48dd12379, please check neutron logs for more information. [ 527.425163] env[63293]: ERROR nova.compute.manager [ 527.425163] env[63293]: Traceback (most recent call last): [ 527.425163] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 527.425163] env[63293]: listener.cb(fileno) [ 527.425163] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.425163] env[63293]: result = function(*args, **kwargs) [ 527.425163] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 527.425163] env[63293]: return func(*args, **kwargs) [ 527.425163] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.425163] env[63293]: raise e [ 527.425163] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.425163] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 527.425163] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.425163] env[63293]: created_port_ids = self._update_ports_for_instance( [ 527.425163] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.425163] env[63293]: with excutils.save_and_reraise_exception(): [ 527.425163] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.425163] env[63293]: self.force_reraise() [ 527.425163] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.425163] env[63293]: raise self.value [ 527.425163] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.425163] env[63293]: updated_port = self._update_port( [ 527.425163] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.425163] env[63293]: _ensure_no_port_binding_failure(port) [ 527.425163] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.425163] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 527.425896] env[63293]: nova.exception.PortBindingFailed: Binding failed for port ead43854-09d6-4e20-ab13-c5e48dd12379, please check neutron logs for more information. [ 527.425896] env[63293]: Removing descriptor: 15 [ 527.425896] env[63293]: ERROR nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ead43854-09d6-4e20-ab13-c5e48dd12379, please check neutron logs for more information. [ 527.425896] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Traceback (most recent call last): [ 527.425896] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 527.425896] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] yield resources [ 527.425896] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 527.425896] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] self.driver.spawn(context, instance, image_meta, [ 527.425896] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 527.425896] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] self._vmops.spawn(context, instance, image_meta, injected_files, [ 527.425896] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 527.425896] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] vm_ref = self.build_virtual_machine(instance, [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] vif_infos = vmwarevif.get_vif_info(self._session, [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] for vif in network_info: [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] return self._sync_wrapper(fn, *args, **kwargs) [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] self.wait() [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] self[:] = self._gt.wait() [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] return self._exit_event.wait() [ 527.426382] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] result = hub.switch() [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] return self.greenlet.switch() [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] result = function(*args, **kwargs) [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] return func(*args, **kwargs) [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] raise e [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] nwinfo = self.network_api.allocate_for_instance( [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.426718] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] created_port_ids = self._update_ports_for_instance( [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] with excutils.save_and_reraise_exception(): [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] self.force_reraise() [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] raise self.value [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] updated_port = self._update_port( [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] _ensure_no_port_binding_failure(port) [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.427129] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] raise exception.PortBindingFailed(port_id=port['id']) [ 527.427432] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] nova.exception.PortBindingFailed: Binding failed for port ead43854-09d6-4e20-ab13-c5e48dd12379, please check neutron logs for more information. [ 527.427432] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] [ 527.427432] env[63293]: INFO nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Terminating instance [ 527.429196] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Acquiring lock "refresh_cache-f9e64134-f233-48a3-afc3-be7765a67c35" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.429399] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Acquired lock "refresh_cache-f9e64134-f233-48a3-afc3-be7765a67c35" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.429565] env[63293]: DEBUG nova.network.neutron [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 527.507866] env[63293]: DEBUG nova.scheduler.client.report [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 527.694795] env[63293]: DEBUG nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 527.722513] env[63293]: DEBUG nova.virt.hardware [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.722997] env[63293]: DEBUG nova.virt.hardware [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.723209] env[63293]: DEBUG nova.virt.hardware [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.723455] env[63293]: DEBUG nova.virt.hardware [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.723740] env[63293]: DEBUG nova.virt.hardware [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.723803] env[63293]: DEBUG nova.virt.hardware [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.724255] env[63293]: DEBUG nova.virt.hardware [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.724466] env[63293]: DEBUG nova.virt.hardware [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.724767] env[63293]: DEBUG nova.virt.hardware [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.724830] env[63293]: DEBUG nova.virt.hardware [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.724973] env[63293]: DEBUG nova.virt.hardware [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.725869] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da168b8-5ba8-43e8-94c0-4f685f379df0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.741673] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0d934a-ca31-44b2-b2df-617a6969191f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.844708] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "db725426-1b67-437c-aea5-cf0da50454a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.844988] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "db725426-1b67-437c-aea5-cf0da50454a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.923472] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327410, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025614} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.923801] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] File moved {{(pid=63293) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 527.924214] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Cleaning up location [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 527.924508] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleting the datastore file [datastore2] vmware_temp/b73c239d-cb78-4d79-b1a9-7124c3ea484b {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 527.924726] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20144be0-1103-4027-ae42-7cb2f1f5aa4a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.932132] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 527.932132] env[63293]: value = "task-1327411" [ 527.932132] env[63293]: _type = "Task" [ 527.932132] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.945112] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327411, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.986375] env[63293]: DEBUG nova.network.neutron [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.014064] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.016422] env[63293]: DEBUG nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 528.022114] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.409s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.022114] env[63293]: INFO nova.compute.claims [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 528.159497] env[63293]: DEBUG nova.network.neutron [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.452661] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327411, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027505} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.452661] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 528.453209] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f7d8ce0-0cef-4327-a402-4f731d0e02c1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.461492] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 528.461492] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52107022-ac5d-86fb-097c-dd09b93f4eee" [ 528.461492] env[63293]: _type = "Task" [ 528.461492] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.472401] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52107022-ac5d-86fb-097c-dd09b93f4eee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.532712] env[63293]: DEBUG nova.compute.utils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 528.537301] env[63293]: DEBUG nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 528.537691] env[63293]: DEBUG nova.network.neutron [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 528.665226] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Releasing lock "refresh_cache-f9e64134-f233-48a3-afc3-be7765a67c35" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.665226] env[63293]: DEBUG nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 528.665226] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 528.665591] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-abefbd8a-1df5-4fe3-98a6-c6f10b1af93d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.678631] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b117d64-4763-48ea-9533-934826a1f7bb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.713826] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f9e64134-f233-48a3-afc3-be7765a67c35 could not be found. [ 528.716024] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 528.716024] env[63293]: INFO nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Took 0.05 seconds to destroy the instance on the hypervisor. [ 528.716024] env[63293]: DEBUG oslo.service.loopingcall [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 528.716024] env[63293]: DEBUG nova.compute.manager [-] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 528.716024] env[63293]: DEBUG nova.network.neutron [-] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 528.743132] env[63293]: DEBUG nova.network.neutron [-] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.782699] env[63293]: DEBUG nova.policy [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4410a50c3dbc4a23aa07aa47c481f1c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bbbd69b9803b4a4e834eafa490241276', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 528.980715] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52107022-ac5d-86fb-097c-dd09b93f4eee, 'name': SearchDatastore_Task, 'duration_secs': 0.010194} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.982841] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.982841] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 528.982841] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b270644-6a25-414c-910e-713cce022975 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.990191] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 528.990191] env[63293]: value = "task-1327412" [ 528.990191] env[63293]: _type = "Task" [ 528.990191] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.999392] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327412, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.040590] env[63293]: DEBUG nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 529.248068] env[63293]: DEBUG nova.network.neutron [-] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.381383] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7c2f4e-44dc-4f77-9888-16b9604c2eee {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.391823] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e633844-215b-4a59-92c5-a419a1bf87ca {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.432209] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b92187f5-8c8b-49cc-b870-6623f2973e33 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.446833] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc95d2a2-4f0d-49d8-9b53-ca34339c693b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.465833] env[63293]: DEBUG nova.compute.provider_tree [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 529.502512] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327412, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.753866] env[63293]: INFO nova.compute.manager [-] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Took 1.04 seconds to deallocate network for instance. [ 529.756388] env[63293]: DEBUG nova.compute.claims [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 529.756543] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.819942] env[63293]: ERROR nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9f614a5-9c36-44bc-8b23-6b8cc5e475d9, please check neutron logs for more information. [ 529.819942] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 529.819942] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.819942] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 529.819942] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.819942] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 529.819942] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.819942] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 529.819942] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.819942] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 529.819942] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.819942] env[63293]: ERROR nova.compute.manager raise self.value [ 529.819942] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.819942] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 529.819942] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.819942] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 529.820437] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.820437] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 529.820437] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9f614a5-9c36-44bc-8b23-6b8cc5e475d9, please check neutron logs for more information. [ 529.820437] env[63293]: ERROR nova.compute.manager [ 529.823018] env[63293]: Traceback (most recent call last): [ 529.823018] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 529.823018] env[63293]: listener.cb(fileno) [ 529.823018] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.823018] env[63293]: result = function(*args, **kwargs) [ 529.823018] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 529.823018] env[63293]: return func(*args, **kwargs) [ 529.823018] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.823018] env[63293]: raise e [ 529.823018] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.823018] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 529.823018] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.823018] env[63293]: created_port_ids = self._update_ports_for_instance( [ 529.823018] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.823018] env[63293]: with excutils.save_and_reraise_exception(): [ 529.823018] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.823018] env[63293]: self.force_reraise() [ 529.823018] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.823018] env[63293]: raise self.value [ 529.823018] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.823018] env[63293]: updated_port = self._update_port( [ 529.823018] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.823018] env[63293]: _ensure_no_port_binding_failure(port) [ 529.823018] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.823018] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 529.823018] env[63293]: nova.exception.PortBindingFailed: Binding failed for port a9f614a5-9c36-44bc-8b23-6b8cc5e475d9, please check neutron logs for more information. [ 529.823018] env[63293]: Removing descriptor: 17 [ 529.823866] env[63293]: ERROR nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9f614a5-9c36-44bc-8b23-6b8cc5e475d9, please check neutron logs for more information. [ 529.823866] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Traceback (most recent call last): [ 529.823866] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 529.823866] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] yield resources [ 529.823866] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 529.823866] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] self.driver.spawn(context, instance, image_meta, [ 529.823866] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 529.823866] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.823866] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.823866] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] vm_ref = self.build_virtual_machine(instance, [ 529.823866] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] for vif in network_info: [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] return self._sync_wrapper(fn, *args, **kwargs) [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] self.wait() [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] self[:] = self._gt.wait() [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] return self._exit_event.wait() [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.824230] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] result = hub.switch() [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] return self.greenlet.switch() [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] result = function(*args, **kwargs) [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] return func(*args, **kwargs) [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] raise e [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] nwinfo = self.network_api.allocate_for_instance( [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] created_port_ids = self._update_ports_for_instance( [ 529.824637] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] with excutils.save_and_reraise_exception(): [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] self.force_reraise() [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] raise self.value [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] updated_port = self._update_port( [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] _ensure_no_port_binding_failure(port) [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] raise exception.PortBindingFailed(port_id=port['id']) [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] nova.exception.PortBindingFailed: Binding failed for port a9f614a5-9c36-44bc-8b23-6b8cc5e475d9, please check neutron logs for more information. [ 529.824987] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] [ 529.825382] env[63293]: INFO nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Terminating instance [ 529.827174] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Acquiring lock "refresh_cache-c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.827174] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Acquired lock "refresh_cache-c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.827174] env[63293]: DEBUG nova.network.neutron [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 529.974859] env[63293]: DEBUG nova.scheduler.client.report [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.987813] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Acquiring lock "76ceec24-eebe-4b32-b77d-f57203881a94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.988120] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Lock "76ceec24-eebe-4b32-b77d-f57203881a94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.005048] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327412, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510614} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.005915] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 530.006448] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 530.006855] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec38f561-b810-4c36-ae1a-116fbaa24769 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.018397] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 530.018397] env[63293]: value = "task-1327413" [ 530.018397] env[63293]: _type = "Task" [ 530.018397] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.028275] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327413, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.061811] env[63293]: DEBUG nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 530.108033] env[63293]: DEBUG nova.network.neutron [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Successfully created port: 2251e186-6de4-4951-9bda-842fc18ba5bf {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 530.125880] env[63293]: DEBUG nova.virt.hardware [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 530.126658] env[63293]: DEBUG nova.virt.hardware [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 530.126658] env[63293]: DEBUG nova.virt.hardware [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 530.126658] env[63293]: DEBUG nova.virt.hardware [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 530.126658] env[63293]: DEBUG nova.virt.hardware [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 530.126658] env[63293]: DEBUG nova.virt.hardware [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 530.127200] env[63293]: DEBUG nova.virt.hardware [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 530.127200] env[63293]: DEBUG nova.virt.hardware [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 530.127200] env[63293]: DEBUG nova.virt.hardware [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 530.127357] env[63293]: DEBUG nova.virt.hardware [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 530.127520] env[63293]: DEBUG nova.virt.hardware [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 530.128406] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4f3d99-3976-4edd-b685-f556fe293be4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.140140] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4475d5-e7f4-427b-a3c4-1aa38c4981b0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.359147] env[63293]: DEBUG nova.network.neutron [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.478438] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.479350] env[63293]: DEBUG nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 530.482866] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.256s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.487125] env[63293]: INFO nova.compute.claims [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 530.492563] env[63293]: DEBUG nova.compute.manager [req-d7097ae4-93a9-499c-bb02-5549526fd77e req-813fe913-f3b0-4204-b852-48aab76a8ffe service nova] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Received event network-vif-deleted-6a6d81e5-3c63-425b-938b-2294443c2fd4 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.492768] env[63293]: DEBUG nova.compute.manager [req-d7097ae4-93a9-499c-bb02-5549526fd77e req-813fe913-f3b0-4204-b852-48aab76a8ffe service nova] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Received event network-changed-ead43854-09d6-4e20-ab13-c5e48dd12379 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.492923] env[63293]: DEBUG nova.compute.manager [req-d7097ae4-93a9-499c-bb02-5549526fd77e req-813fe913-f3b0-4204-b852-48aab76a8ffe service nova] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Refreshing instance network info cache due to event network-changed-ead43854-09d6-4e20-ab13-c5e48dd12379. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 530.493142] env[63293]: DEBUG oslo_concurrency.lockutils [req-d7097ae4-93a9-499c-bb02-5549526fd77e req-813fe913-f3b0-4204-b852-48aab76a8ffe service nova] Acquiring lock "refresh_cache-f9e64134-f233-48a3-afc3-be7765a67c35" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.493303] env[63293]: DEBUG oslo_concurrency.lockutils [req-d7097ae4-93a9-499c-bb02-5549526fd77e req-813fe913-f3b0-4204-b852-48aab76a8ffe service nova] Acquired lock "refresh_cache-f9e64134-f233-48a3-afc3-be7765a67c35" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.493463] env[63293]: DEBUG nova.network.neutron [req-d7097ae4-93a9-499c-bb02-5549526fd77e req-813fe913-f3b0-4204-b852-48aab76a8ffe service nova] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Refreshing network info cache for port ead43854-09d6-4e20-ab13-c5e48dd12379 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 530.496385] env[63293]: DEBUG nova.network.neutron [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.536435] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327413, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078562} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.536732] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 530.539156] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c26923-d1d6-4a35-b9d4-067897219eea {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.567293] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 530.567293] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e049dd9-e6c5-427e-951e-e9e1f2d88993 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.589428] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 530.589428] env[63293]: value = "task-1327414" [ 530.589428] env[63293]: _type = "Task" [ 530.589428] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.597850] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327414, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.999591] env[63293]: DEBUG nova.compute.utils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 531.004149] env[63293]: DEBUG nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 531.004740] env[63293]: DEBUG nova.network.neutron [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 531.007890] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Releasing lock "refresh_cache-c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.008354] env[63293]: DEBUG nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 531.009383] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 531.012281] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fc62f3e1-2523-4a25-862e-ec57799c09c2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.018936] env[63293]: DEBUG nova.compute.manager [req-331848b7-5645-4ea7-9dbc-6745edba4a10 req-48a8ca91-80fe-4aa2-81f0-ea379b6ab4b4 service nova] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Received event network-changed-a9f614a5-9c36-44bc-8b23-6b8cc5e475d9 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 531.019198] env[63293]: DEBUG nova.compute.manager [req-331848b7-5645-4ea7-9dbc-6745edba4a10 req-48a8ca91-80fe-4aa2-81f0-ea379b6ab4b4 service nova] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Refreshing instance network info cache due to event network-changed-a9f614a5-9c36-44bc-8b23-6b8cc5e475d9. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 531.019661] env[63293]: DEBUG oslo_concurrency.lockutils [req-331848b7-5645-4ea7-9dbc-6745edba4a10 req-48a8ca91-80fe-4aa2-81f0-ea379b6ab4b4 service nova] Acquiring lock "refresh_cache-c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.019957] env[63293]: DEBUG oslo_concurrency.lockutils [req-331848b7-5645-4ea7-9dbc-6745edba4a10 req-48a8ca91-80fe-4aa2-81f0-ea379b6ab4b4 service nova] Acquired lock "refresh_cache-c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.020181] env[63293]: DEBUG nova.network.neutron [req-331848b7-5645-4ea7-9dbc-6745edba4a10 req-48a8ca91-80fe-4aa2-81f0-ea379b6ab4b4 service nova] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Refreshing network info cache for port a9f614a5-9c36-44bc-8b23-6b8cc5e475d9 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 531.027622] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a2ee46-3dba-449d-a0c0-cd24ab537277 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.047317] env[63293]: DEBUG nova.network.neutron [req-d7097ae4-93a9-499c-bb02-5549526fd77e req-813fe913-f3b0-4204-b852-48aab76a8ffe service nova] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.054670] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4 could not be found. [ 531.055266] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 531.055507] env[63293]: INFO nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 531.056140] env[63293]: DEBUG oslo.service.loopingcall [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 531.056667] env[63293]: DEBUG nova.compute.manager [-] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 531.056965] env[63293]: DEBUG nova.network.neutron [-] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 531.090067] env[63293]: DEBUG nova.network.neutron [-] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.101687] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327414, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.165266] env[63293]: DEBUG nova.policy [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03a004f9d0764e60b3b3b76ea99fb0b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6285f4d2ba024795bb144c8ff5d583e2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 531.227123] env[63293]: DEBUG nova.network.neutron [req-d7097ae4-93a9-499c-bb02-5549526fd77e req-813fe913-f3b0-4204-b852-48aab76a8ffe service nova] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.513735] env[63293]: DEBUG nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 531.602636] env[63293]: DEBUG nova.network.neutron [-] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.616153] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327414, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.629937] env[63293]: DEBUG nova.network.neutron [req-331848b7-5645-4ea7-9dbc-6745edba4a10 req-48a8ca91-80fe-4aa2-81f0-ea379b6ab4b4 service nova] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.731034] env[63293]: DEBUG oslo_concurrency.lockutils [req-d7097ae4-93a9-499c-bb02-5549526fd77e req-813fe913-f3b0-4204-b852-48aab76a8ffe service nova] Releasing lock "refresh_cache-f9e64134-f233-48a3-afc3-be7765a67c35" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.839087] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd466737-1d23-4710-957a-6ec27b3c1cf4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.847157] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95247e99-2878-4111-aa69-e83c20526ae9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.882556] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8e26e9-cef0-415c-9b5f-f8b6fc2a391f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.890363] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35fc3de-f02d-41e7-9b5c-3c6de14a8a97 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.904633] env[63293]: DEBUG nova.compute.provider_tree [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.926733] env[63293]: DEBUG nova.network.neutron [req-331848b7-5645-4ea7-9dbc-6745edba4a10 req-48a8ca91-80fe-4aa2-81f0-ea379b6ab4b4 service nova] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.100776] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327414, 'name': ReconfigVM_Task, 'duration_secs': 1.263247} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.100971] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Reconfigured VM instance instance-00000002 to attach disk [datastore2] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 532.102941] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09561f02-d23a-4027-9d56-be0e818457ec {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.109761] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 532.109761] env[63293]: value = "task-1327415" [ 532.109761] env[63293]: _type = "Task" [ 532.109761] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.117097] env[63293]: INFO nova.compute.manager [-] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Took 1.06 seconds to deallocate network for instance. [ 532.126923] env[63293]: DEBUG nova.compute.claims [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 532.127081] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.127235] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327415, 'name': Rename_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.185568] env[63293]: DEBUG nova.network.neutron [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Successfully created port: 2841f92c-3f53-4e90-a6c8-0e794b20954a {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.220665] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Acquiring lock "25f0de66-914f-46b9-a15b-d936df5b3ab4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.221118] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Lock "25f0de66-914f-46b9-a15b-d936df5b3ab4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.410235] env[63293]: DEBUG nova.scheduler.client.report [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 532.429508] env[63293]: DEBUG oslo_concurrency.lockutils [req-331848b7-5645-4ea7-9dbc-6745edba4a10 req-48a8ca91-80fe-4aa2-81f0-ea379b6ab4b4 service nova] Releasing lock "refresh_cache-c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.527855] env[63293]: DEBUG nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 532.564934] env[63293]: DEBUG nova.virt.hardware [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 532.565179] env[63293]: DEBUG nova.virt.hardware [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 532.565332] env[63293]: DEBUG nova.virt.hardware [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 532.565504] env[63293]: DEBUG nova.virt.hardware [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 532.565647] env[63293]: DEBUG nova.virt.hardware [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 532.565866] env[63293]: DEBUG nova.virt.hardware [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 532.566028] env[63293]: DEBUG nova.virt.hardware [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 532.566173] env[63293]: DEBUG nova.virt.hardware [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 532.566344] env[63293]: DEBUG nova.virt.hardware [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 532.566811] env[63293]: DEBUG nova.virt.hardware [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 532.566811] env[63293]: DEBUG nova.virt.hardware [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 532.567710] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a052759d-359a-4735-9f06-c29a345e6758 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.578328] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c4b3b0-c219-453e-ba81-40c879d88245 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.620334] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327415, 'name': Rename_Task, 'duration_secs': 0.133821} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.622458] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 532.622771] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe8ea19b-6366-4ce0-a1ef-116ebef50258 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.632651] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 532.632651] env[63293]: value = "task-1327416" [ 532.632651] env[63293]: _type = "Task" [ 532.632651] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.638616] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.848067] env[63293]: ERROR nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2251e186-6de4-4951-9bda-842fc18ba5bf, please check neutron logs for more information. [ 532.848067] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 532.848067] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.848067] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 532.848067] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.848067] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 532.848067] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.848067] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 532.848067] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.848067] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 532.848067] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.848067] env[63293]: ERROR nova.compute.manager raise self.value [ 532.848067] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.848067] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 532.848067] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.848067] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 532.848920] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.848920] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 532.848920] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2251e186-6de4-4951-9bda-842fc18ba5bf, please check neutron logs for more information. [ 532.848920] env[63293]: ERROR nova.compute.manager [ 532.848920] env[63293]: Traceback (most recent call last): [ 532.848920] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 532.848920] env[63293]: listener.cb(fileno) [ 532.848920] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.848920] env[63293]: result = function(*args, **kwargs) [ 532.848920] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 532.848920] env[63293]: return func(*args, **kwargs) [ 532.848920] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.848920] env[63293]: raise e [ 532.848920] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.848920] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 532.848920] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.848920] env[63293]: created_port_ids = self._update_ports_for_instance( [ 532.848920] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.848920] env[63293]: with excutils.save_and_reraise_exception(): [ 532.848920] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.848920] env[63293]: self.force_reraise() [ 532.848920] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.848920] env[63293]: raise self.value [ 532.848920] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.848920] env[63293]: updated_port = self._update_port( [ 532.848920] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.848920] env[63293]: _ensure_no_port_binding_failure(port) [ 532.848920] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.848920] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 532.849677] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 2251e186-6de4-4951-9bda-842fc18ba5bf, please check neutron logs for more information. [ 532.849677] env[63293]: Removing descriptor: 15 [ 532.849677] env[63293]: ERROR nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2251e186-6de4-4951-9bda-842fc18ba5bf, please check neutron logs for more information. [ 532.849677] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] Traceback (most recent call last): [ 532.849677] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 532.849677] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] yield resources [ 532.849677] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 532.849677] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] self.driver.spawn(context, instance, image_meta, [ 532.849677] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 532.849677] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.849677] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.849677] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] vm_ref = self.build_virtual_machine(instance, [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] for vif in network_info: [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] return self._sync_wrapper(fn, *args, **kwargs) [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] self.wait() [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] self[:] = self._gt.wait() [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] return self._exit_event.wait() [ 532.850019] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] result = hub.switch() [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] return self.greenlet.switch() [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] result = function(*args, **kwargs) [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] return func(*args, **kwargs) [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] raise e [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] nwinfo = self.network_api.allocate_for_instance( [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.850364] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] created_port_ids = self._update_ports_for_instance( [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] with excutils.save_and_reraise_exception(): [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] self.force_reraise() [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] raise self.value [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] updated_port = self._update_port( [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] _ensure_no_port_binding_failure(port) [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.850708] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] raise exception.PortBindingFailed(port_id=port['id']) [ 532.851055] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] nova.exception.PortBindingFailed: Binding failed for port 2251e186-6de4-4951-9bda-842fc18ba5bf, please check neutron logs for more information. [ 532.851055] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] [ 532.851055] env[63293]: INFO nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Terminating instance [ 532.852394] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Acquiring lock "refresh_cache-aa4611e2-6254-4983-8565-899aea616726" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.852456] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Acquired lock "refresh_cache-aa4611e2-6254-4983-8565-899aea616726" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.853212] env[63293]: DEBUG nova.network.neutron [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 532.919933] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.920493] env[63293]: DEBUG nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 532.927142] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.946s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.930598] env[63293]: INFO nova.compute.claims [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 533.018767] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Acquiring lock "c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.019012] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Lock "c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.140164] env[63293]: DEBUG oslo_vmware.api [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327416, 'name': PowerOnVM_Task, 'duration_secs': 0.441707} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.140424] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 533.140594] env[63293]: INFO nova.compute.manager [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Took 10.04 seconds to spawn the instance on the hypervisor. [ 533.140834] env[63293]: DEBUG nova.compute.manager [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 533.141675] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22df029b-fe83-43fd-a48f-67ede02db069 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.383069] env[63293]: DEBUG nova.network.neutron [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.439791] env[63293]: DEBUG nova.compute.utils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 533.449555] env[63293]: DEBUG nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 533.449555] env[63293]: DEBUG nova.network.neutron [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 533.495830] env[63293]: DEBUG nova.network.neutron [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.543957] env[63293]: DEBUG nova.policy [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3bb00827293c4b74a6f86f8db3a09cfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9169d8b45ed148b8935d1f8ca96ac3e0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 533.566531] env[63293]: DEBUG nova.compute.manager [req-10d0595e-000e-434a-bc96-17b23ddbffc6 req-83ef1177-b32a-45fe-8bec-16c9e327c4a3 service nova] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Received event network-vif-deleted-ead43854-09d6-4e20-ab13-c5e48dd12379 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.566722] env[63293]: DEBUG nova.compute.manager [req-10d0595e-000e-434a-bc96-17b23ddbffc6 req-83ef1177-b32a-45fe-8bec-16c9e327c4a3 service nova] [instance: aa4611e2-6254-4983-8565-899aea616726] Received event network-changed-2251e186-6de4-4951-9bda-842fc18ba5bf {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.566876] env[63293]: DEBUG nova.compute.manager [req-10d0595e-000e-434a-bc96-17b23ddbffc6 req-83ef1177-b32a-45fe-8bec-16c9e327c4a3 service nova] [instance: aa4611e2-6254-4983-8565-899aea616726] Refreshing instance network info cache due to event network-changed-2251e186-6de4-4951-9bda-842fc18ba5bf. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 533.567072] env[63293]: DEBUG oslo_concurrency.lockutils [req-10d0595e-000e-434a-bc96-17b23ddbffc6 req-83ef1177-b32a-45fe-8bec-16c9e327c4a3 service nova] Acquiring lock "refresh_cache-aa4611e2-6254-4983-8565-899aea616726" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.664305] env[63293]: INFO nova.compute.manager [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Took 15.56 seconds to build instance. [ 533.803046] env[63293]: DEBUG nova.compute.manager [req-aa660ee7-aef0-42aa-a649-231511e95ed8 req-f6eb5c97-9d75-4b39-9a98-100db05d3a34 service nova] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Received event network-vif-deleted-a9f614a5-9c36-44bc-8b23-6b8cc5e475d9 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.944969] env[63293]: DEBUG nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 534.005994] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Releasing lock "refresh_cache-aa4611e2-6254-4983-8565-899aea616726" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.005994] env[63293]: DEBUG nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 534.005994] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 534.005994] env[63293]: DEBUG oslo_concurrency.lockutils [req-10d0595e-000e-434a-bc96-17b23ddbffc6 req-83ef1177-b32a-45fe-8bec-16c9e327c4a3 service nova] Acquired lock "refresh_cache-aa4611e2-6254-4983-8565-899aea616726" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.005994] env[63293]: DEBUG nova.network.neutron [req-10d0595e-000e-434a-bc96-17b23ddbffc6 req-83ef1177-b32a-45fe-8bec-16c9e327c4a3 service nova] [instance: aa4611e2-6254-4983-8565-899aea616726] Refreshing network info cache for port 2251e186-6de4-4951-9bda-842fc18ba5bf {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 534.006617] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86879194-2eda-43a5-a15a-a3ef5d204d3c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.023954] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d118a7-9e28-4ae6-836e-df9a3e01ba97 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.054960] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aa4611e2-6254-4983-8565-899aea616726 could not be found. [ 534.055238] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 534.055417] env[63293]: INFO nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Took 0.05 seconds to destroy the instance on the hypervisor. [ 534.055660] env[63293]: DEBUG oslo.service.loopingcall [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.061469] env[63293]: DEBUG nova.compute.manager [-] [instance: aa4611e2-6254-4983-8565-899aea616726] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.061580] env[63293]: DEBUG nova.network.neutron [-] [instance: aa4611e2-6254-4983-8565-899aea616726] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 534.096947] env[63293]: DEBUG nova.network.neutron [-] [instance: aa4611e2-6254-4983-8565-899aea616726] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.167190] env[63293]: DEBUG oslo_concurrency.lockutils [None req-661320b0-22a5-485e-a2be-3be01457fc54 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "e80209b3-273c-4045-ad8a-2a3c4c48200a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.072s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.362672] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42144d6e-bb5a-48f2-9173-71ead6ebc999 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.371822] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de6ef3a-fcd3-45d7-8479-88704067efd4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.410262] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35263416-c071-4d81-bfe7-1fc8ab0844f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.416914] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b47d4a6-6603-421a-8ec6-cdedc16cc2a3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.431606] env[63293]: DEBUG nova.compute.provider_tree [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.435808] env[63293]: ERROR nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2841f92c-3f53-4e90-a6c8-0e794b20954a, please check neutron logs for more information. [ 534.435808] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.435808] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.435808] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.435808] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.435808] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.435808] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.435808] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.435808] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.435808] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 534.435808] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.435808] env[63293]: ERROR nova.compute.manager raise self.value [ 534.435808] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.435808] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.435808] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.435808] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.437084] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.437084] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.437084] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2841f92c-3f53-4e90-a6c8-0e794b20954a, please check neutron logs for more information. [ 534.437084] env[63293]: ERROR nova.compute.manager [ 534.437084] env[63293]: Traceback (most recent call last): [ 534.437084] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.437084] env[63293]: listener.cb(fileno) [ 534.437084] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.437084] env[63293]: result = function(*args, **kwargs) [ 534.437084] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.437084] env[63293]: return func(*args, **kwargs) [ 534.437084] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.437084] env[63293]: raise e [ 534.437084] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.437084] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 534.437084] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.437084] env[63293]: created_port_ids = self._update_ports_for_instance( [ 534.437084] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.437084] env[63293]: with excutils.save_and_reraise_exception(): [ 534.437084] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.437084] env[63293]: self.force_reraise() [ 534.437084] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.437084] env[63293]: raise self.value [ 534.437084] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.437084] env[63293]: updated_port = self._update_port( [ 534.437084] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.437084] env[63293]: _ensure_no_port_binding_failure(port) [ 534.437084] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.437084] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.439195] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 2841f92c-3f53-4e90-a6c8-0e794b20954a, please check neutron logs for more information. [ 534.439195] env[63293]: Removing descriptor: 17 [ 534.439195] env[63293]: ERROR nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2841f92c-3f53-4e90-a6c8-0e794b20954a, please check neutron logs for more information. [ 534.439195] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Traceback (most recent call last): [ 534.439195] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 534.439195] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] yield resources [ 534.439195] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.439195] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] self.driver.spawn(context, instance, image_meta, [ 534.439195] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 534.439195] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.439195] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.439195] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] vm_ref = self.build_virtual_machine(instance, [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] for vif in network_info: [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] return self._sync_wrapper(fn, *args, **kwargs) [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] self.wait() [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] self[:] = self._gt.wait() [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] return self._exit_event.wait() [ 534.439536] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] result = hub.switch() [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] return self.greenlet.switch() [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] result = function(*args, **kwargs) [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] return func(*args, **kwargs) [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] raise e [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] nwinfo = self.network_api.allocate_for_instance( [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.440960] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] created_port_ids = self._update_ports_for_instance( [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] with excutils.save_and_reraise_exception(): [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] self.force_reraise() [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] raise self.value [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] updated_port = self._update_port( [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] _ensure_no_port_binding_failure(port) [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.441319] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] raise exception.PortBindingFailed(port_id=port['id']) [ 534.441639] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] nova.exception.PortBindingFailed: Binding failed for port 2841f92c-3f53-4e90-a6c8-0e794b20954a, please check neutron logs for more information. [ 534.441639] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] [ 534.441639] env[63293]: INFO nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Terminating instance [ 534.442951] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquiring lock "refresh_cache-c9221012-9f13-4881-9a0a-1b9c1c84d910" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.443353] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquired lock "refresh_cache-c9221012-9f13-4881-9a0a-1b9c1c84d910" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.443353] env[63293]: DEBUG nova.network.neutron [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 534.470674] env[63293]: DEBUG nova.network.neutron [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Successfully created port: 005e83b3-4351-48e5-aefc-c6620e43ea0b {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 534.558080] env[63293]: DEBUG nova.network.neutron [req-10d0595e-000e-434a-bc96-17b23ddbffc6 req-83ef1177-b32a-45fe-8bec-16c9e327c4a3 service nova] [instance: aa4611e2-6254-4983-8565-899aea616726] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.600678] env[63293]: DEBUG nova.network.neutron [-] [instance: aa4611e2-6254-4983-8565-899aea616726] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.682323] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 534.745082] env[63293]: DEBUG nova.network.neutron [req-10d0595e-000e-434a-bc96-17b23ddbffc6 req-83ef1177-b32a-45fe-8bec-16c9e327c4a3 service nova] [instance: aa4611e2-6254-4983-8565-899aea616726] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.938741] env[63293]: DEBUG nova.scheduler.client.report [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 534.958264] env[63293]: DEBUG nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 534.987335] env[63293]: DEBUG nova.virt.hardware [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 534.987573] env[63293]: DEBUG nova.virt.hardware [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 534.987742] env[63293]: DEBUG nova.virt.hardware [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.987897] env[63293]: DEBUG nova.virt.hardware [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 534.988065] env[63293]: DEBUG nova.virt.hardware [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.988220] env[63293]: DEBUG nova.virt.hardware [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 534.988438] env[63293]: DEBUG nova.virt.hardware [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 534.988606] env[63293]: DEBUG nova.virt.hardware [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.988765] env[63293]: DEBUG nova.virt.hardware [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.988922] env[63293]: DEBUG nova.virt.hardware [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.989113] env[63293]: DEBUG nova.virt.hardware [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.989952] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40cca742-fe6c-4bf2-8121-1b8adf903119 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.998845] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983ffd84-3931-4c6f-a62f-e86ee7756d35 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.022140] env[63293]: DEBUG nova.network.neutron [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.107131] env[63293]: INFO nova.compute.manager [-] [instance: aa4611e2-6254-4983-8565-899aea616726] Took 1.04 seconds to deallocate network for instance. [ 535.107720] env[63293]: DEBUG nova.compute.claims [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 535.107875] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.201836] env[63293]: DEBUG nova.network.neutron [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.209849] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.249624] env[63293]: DEBUG oslo_concurrency.lockutils [req-10d0595e-000e-434a-bc96-17b23ddbffc6 req-83ef1177-b32a-45fe-8bec-16c9e327c4a3 service nova] Releasing lock "refresh_cache-aa4611e2-6254-4983-8565-899aea616726" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.445211] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.445985] env[63293]: DEBUG nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 535.449907] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.505s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.452132] env[63293]: INFO nova.compute.claims [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 535.707772] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Releasing lock "refresh_cache-c9221012-9f13-4881-9a0a-1b9c1c84d910" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.708672] env[63293]: DEBUG nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.708672] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 535.709941] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15de6d3f-1732-48ea-9e82-928c61d24534 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.722195] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a329e491-8b0e-4547-841b-80d47b6fb9b0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.752024] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c9221012-9f13-4881-9a0a-1b9c1c84d910 could not be found. [ 535.752690] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 535.752690] env[63293]: INFO nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Took 0.04 seconds to destroy the instance on the hypervisor. [ 535.752884] env[63293]: DEBUG oslo.service.loopingcall [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.753012] env[63293]: DEBUG nova.compute.manager [-] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.753135] env[63293]: DEBUG nova.network.neutron [-] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 535.804991] env[63293]: DEBUG nova.network.neutron [-] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.957606] env[63293]: DEBUG nova.compute.utils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 535.962065] env[63293]: DEBUG nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 535.962065] env[63293]: DEBUG nova.network.neutron [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 536.158648] env[63293]: DEBUG nova.policy [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '08823c8ad18a4c7b894289e4fafb879e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7eccd94f77f8438598cdb8d357442819', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 536.311335] env[63293]: DEBUG nova.network.neutron [-] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.462920] env[63293]: DEBUG nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 536.816785] env[63293]: INFO nova.compute.manager [-] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Took 1.06 seconds to deallocate network for instance. [ 536.824820] env[63293]: DEBUG nova.compute.claims [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.825060] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.903053] env[63293]: DEBUG nova.compute.manager [req-29ae631d-7449-4d4d-be2d-b06ac6b52c61 req-53020e03-1bac-4c1d-88a5-ea978a80b101 service nova] [instance: aa4611e2-6254-4983-8565-899aea616726] Received event network-vif-deleted-2251e186-6de4-4951-9bda-842fc18ba5bf {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 536.912339] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87fa8120-de3c-4d7a-b482-b1e389e3d33f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.922059] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871a84a0-2c5c-43e4-bf84-1a013a4bbd63 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.958369] env[63293]: DEBUG nova.compute.manager [req-f409d216-e18a-4630-a049-fccd7e8d2df5 req-f9689554-ce44-42fd-8231-de083e87e3fa service nova] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Received event network-changed-2841f92c-3f53-4e90-a6c8-0e794b20954a {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 536.958584] env[63293]: DEBUG nova.compute.manager [req-f409d216-e18a-4630-a049-fccd7e8d2df5 req-f9689554-ce44-42fd-8231-de083e87e3fa service nova] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Refreshing instance network info cache due to event network-changed-2841f92c-3f53-4e90-a6c8-0e794b20954a. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 536.958794] env[63293]: DEBUG oslo_concurrency.lockutils [req-f409d216-e18a-4630-a049-fccd7e8d2df5 req-f9689554-ce44-42fd-8231-de083e87e3fa service nova] Acquiring lock "refresh_cache-c9221012-9f13-4881-9a0a-1b9c1c84d910" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.958931] env[63293]: DEBUG oslo_concurrency.lockutils [req-f409d216-e18a-4630-a049-fccd7e8d2df5 req-f9689554-ce44-42fd-8231-de083e87e3fa service nova] Acquired lock "refresh_cache-c9221012-9f13-4881-9a0a-1b9c1c84d910" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.959102] env[63293]: DEBUG nova.network.neutron [req-f409d216-e18a-4630-a049-fccd7e8d2df5 req-f9689554-ce44-42fd-8231-de083e87e3fa service nova] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Refreshing network info cache for port 2841f92c-3f53-4e90-a6c8-0e794b20954a {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 536.960614] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e2b5e2-7da4-4f21-b385-1fd91bf0bd5f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.970162] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3e2c1f-fe26-4238-9a84-3e895ec20951 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.988774] env[63293]: DEBUG nova.compute.provider_tree [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.053138] env[63293]: DEBUG nova.network.neutron [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Successfully created port: 9c1fae21-16d8-4377-8f41-e9c9154ee27f {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 537.151405] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Acquiring lock "a236be17-8965-4785-834e-9ce22eeb75eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.151405] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Lock "a236be17-8965-4785-834e-9ce22eeb75eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.303657] env[63293]: INFO nova.compute.manager [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Rebuilding instance [ 537.360052] env[63293]: DEBUG nova.compute.manager [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 537.360870] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3914c028-59e4-4185-9da1-4b6bfc5e6afb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.508740] env[63293]: DEBUG nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 537.512030] env[63293]: DEBUG nova.scheduler.client.report [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 537.524249] env[63293]: DEBUG nova.network.neutron [req-f409d216-e18a-4630-a049-fccd7e8d2df5 req-f9689554-ce44-42fd-8231-de083e87e3fa service nova] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.551745] env[63293]: DEBUG nova.virt.hardware [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 537.552039] env[63293]: DEBUG nova.virt.hardware [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 537.552421] env[63293]: DEBUG nova.virt.hardware [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 537.552878] env[63293]: DEBUG nova.virt.hardware [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 537.552878] env[63293]: DEBUG nova.virt.hardware [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 537.553990] env[63293]: DEBUG nova.virt.hardware [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 537.553990] env[63293]: DEBUG nova.virt.hardware [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 537.553990] env[63293]: DEBUG nova.virt.hardware [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 537.553990] env[63293]: DEBUG nova.virt.hardware [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 537.553990] env[63293]: DEBUG nova.virt.hardware [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 537.554337] env[63293]: DEBUG nova.virt.hardware [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 537.555922] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3799d2-9617-4900-a1b6-d459df1ac7fe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.565529] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d210b4c6-1ac6-4ac8-b568-47b86eb32ee6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.736408] env[63293]: DEBUG nova.network.neutron [req-f409d216-e18a-4630-a049-fccd7e8d2df5 req-f9689554-ce44-42fd-8231-de083e87e3fa service nova] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.876965] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 537.877323] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59409d0a-9acf-4609-8500-c184673212da {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.885634] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 537.885634] env[63293]: value = "task-1327417" [ 537.885634] env[63293]: _type = "Task" [ 537.885634] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.895649] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327417, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.019809] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.019809] env[63293]: DEBUG nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 538.024763] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.235s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.028820] env[63293]: INFO nova.compute.claims [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 538.149381] env[63293]: ERROR nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 005e83b3-4351-48e5-aefc-c6620e43ea0b, please check neutron logs for more information. [ 538.149381] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 538.149381] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.149381] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 538.149381] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.149381] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 538.149381] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.149381] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 538.149381] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.149381] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 538.149381] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.149381] env[63293]: ERROR nova.compute.manager raise self.value [ 538.149381] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.149381] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 538.149381] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.149381] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 538.149841] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.149841] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 538.149841] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 005e83b3-4351-48e5-aefc-c6620e43ea0b, please check neutron logs for more information. [ 538.149841] env[63293]: ERROR nova.compute.manager [ 538.149841] env[63293]: Traceback (most recent call last): [ 538.149841] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 538.149841] env[63293]: listener.cb(fileno) [ 538.149841] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.149841] env[63293]: result = function(*args, **kwargs) [ 538.149841] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 538.149841] env[63293]: return func(*args, **kwargs) [ 538.149841] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.149841] env[63293]: raise e [ 538.149841] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.149841] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 538.149841] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.149841] env[63293]: created_port_ids = self._update_ports_for_instance( [ 538.149841] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.149841] env[63293]: with excutils.save_and_reraise_exception(): [ 538.149841] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.149841] env[63293]: self.force_reraise() [ 538.149841] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.149841] env[63293]: raise self.value [ 538.149841] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.149841] env[63293]: updated_port = self._update_port( [ 538.149841] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.149841] env[63293]: _ensure_no_port_binding_failure(port) [ 538.149841] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.149841] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 538.150724] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 005e83b3-4351-48e5-aefc-c6620e43ea0b, please check neutron logs for more information. [ 538.150724] env[63293]: Removing descriptor: 18 [ 538.150724] env[63293]: ERROR nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 005e83b3-4351-48e5-aefc-c6620e43ea0b, please check neutron logs for more information. [ 538.150724] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Traceback (most recent call last): [ 538.150724] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 538.150724] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] yield resources [ 538.150724] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 538.150724] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] self.driver.spawn(context, instance, image_meta, [ 538.150724] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 538.150724] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.150724] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.150724] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] vm_ref = self.build_virtual_machine(instance, [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] for vif in network_info: [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] return self._sync_wrapper(fn, *args, **kwargs) [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] self.wait() [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] self[:] = self._gt.wait() [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] return self._exit_event.wait() [ 538.151072] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] result = hub.switch() [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] return self.greenlet.switch() [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] result = function(*args, **kwargs) [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] return func(*args, **kwargs) [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] raise e [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] nwinfo = self.network_api.allocate_for_instance( [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.151620] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] created_port_ids = self._update_ports_for_instance( [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] with excutils.save_and_reraise_exception(): [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] self.force_reraise() [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] raise self.value [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] updated_port = self._update_port( [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] _ensure_no_port_binding_failure(port) [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.151961] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] raise exception.PortBindingFailed(port_id=port['id']) [ 538.152272] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] nova.exception.PortBindingFailed: Binding failed for port 005e83b3-4351-48e5-aefc-c6620e43ea0b, please check neutron logs for more information. [ 538.152272] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] [ 538.152272] env[63293]: INFO nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Terminating instance [ 538.156283] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "refresh_cache-f255ca12-7110-4f73-a6ff-d6bcc122486b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.156283] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquired lock "refresh_cache-f255ca12-7110-4f73-a6ff-d6bcc122486b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.156283] env[63293]: DEBUG nova.network.neutron [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 538.242630] env[63293]: DEBUG oslo_concurrency.lockutils [req-f409d216-e18a-4630-a049-fccd7e8d2df5 req-f9689554-ce44-42fd-8231-de083e87e3fa service nova] Releasing lock "refresh_cache-c9221012-9f13-4881-9a0a-1b9c1c84d910" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.242986] env[63293]: DEBUG nova.compute.manager [req-f409d216-e18a-4630-a049-fccd7e8d2df5 req-f9689554-ce44-42fd-8231-de083e87e3fa service nova] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Received event network-vif-deleted-2841f92c-3f53-4e90-a6c8-0e794b20954a {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.398490] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327417, 'name': PowerOffVM_Task, 'duration_secs': 0.137143} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.398774] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 538.398993] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 538.399804] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd21b3dc-35d4-48be-8b03-a71d6a469ffa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.406927] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 538.407207] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6a99228-58bb-4420-8dd8-226540f54c1f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.438740] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 538.442527] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 538.442527] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleting the datastore file [datastore2] e80209b3-273c-4045-ad8a-2a3c4c48200a {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 538.442527] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad30056e-15eb-420c-89d3-a4f0481accd8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.449802] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 538.449802] env[63293]: value = "task-1327419" [ 538.449802] env[63293]: _type = "Task" [ 538.449802] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.460167] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327419, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.533958] env[63293]: DEBUG nova.compute.utils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 538.535467] env[63293]: DEBUG nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 538.535688] env[63293]: DEBUG nova.network.neutron [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 538.648045] env[63293]: DEBUG nova.policy [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b905d104776482cb262ee41c0638433', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '29986647349c4ea6a7f0bc878d5161b0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 538.702025] env[63293]: DEBUG nova.network.neutron [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.823243] env[63293]: DEBUG nova.network.neutron [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.965298] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327419, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112504} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.965649] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 538.966156] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 538.966311] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.043332] env[63293]: DEBUG nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 539.247186] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Acquiring lock "a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.247418] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Lock "a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.325750] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Releasing lock "refresh_cache-f255ca12-7110-4f73-a6ff-d6bcc122486b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.326248] env[63293]: DEBUG nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 539.326515] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.329455] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e7cef90-a29e-4524-9c95-d468f5c605b1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.332489] env[63293]: DEBUG nova.network.neutron [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Successfully created port: f5b367dd-acd0-41f7-81bf-cfdcdafe0e82 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 539.341907] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af1d6b0-d749-4f6d-ab03-4ce715bea6c2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.375026] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f255ca12-7110-4f73-a6ff-d6bcc122486b could not be found. [ 539.375026] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.376969] env[63293]: INFO nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 539.376969] env[63293]: DEBUG oslo.service.loopingcall [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.379233] env[63293]: DEBUG nova.compute.manager [-] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.379335] env[63293]: DEBUG nova.network.neutron [-] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 539.419515] env[63293]: DEBUG nova.network.neutron [-] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.507309] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-640b878c-8695-4f7b-b29b-5cecec9f8e9c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.518637] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba2938a-45f3-43fe-aeed-36b9582012d5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.559382] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc530e1-c9be-4fd2-b090-eb682989ebfc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.571107] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fdb91f-ec29-42a8-9543-ac49c0b8f5fa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.590280] env[63293]: DEBUG nova.compute.provider_tree [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.627514] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquiring lock "b2055187-edd8-4a8c-9811-183cf7ba3b39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.627514] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Lock "b2055187-edd8-4a8c-9811-183cf7ba3b39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.926120] env[63293]: DEBUG nova.network.neutron [-] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.005135] env[63293]: DEBUG nova.virt.hardware [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 540.005135] env[63293]: DEBUG nova.virt.hardware [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 540.005135] env[63293]: DEBUG nova.virt.hardware [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 540.005135] env[63293]: DEBUG nova.virt.hardware [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 540.005559] env[63293]: DEBUG nova.virt.hardware [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 540.005559] env[63293]: DEBUG nova.virt.hardware [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 540.005559] env[63293]: DEBUG nova.virt.hardware [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 540.005559] env[63293]: DEBUG nova.virt.hardware [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 540.005559] env[63293]: DEBUG nova.virt.hardware [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 540.006032] env[63293]: DEBUG nova.virt.hardware [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 540.006816] env[63293]: DEBUG nova.virt.hardware [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 540.007323] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c78056-558a-430d-ade1-7b292ab671e7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.018350] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cece184-8e9f-41e3-92f6-f33ae5647b98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.039857] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 540.046416] env[63293]: DEBUG oslo.service.loopingcall [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.046605] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 540.046873] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db729459-4910-4c7c-9e3e-22281c890e65 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.063947] env[63293]: DEBUG nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 540.068182] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 540.068182] env[63293]: value = "task-1327420" [ 540.068182] env[63293]: _type = "Task" [ 540.068182] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.084870] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327420, 'name': CreateVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.090163] env[63293]: ERROR nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9c1fae21-16d8-4377-8f41-e9c9154ee27f, please check neutron logs for more information. [ 540.090163] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.090163] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.090163] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.090163] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.090163] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.090163] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.090163] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.090163] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.090163] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 540.090163] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.090163] env[63293]: ERROR nova.compute.manager raise self.value [ 540.090163] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.090163] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.090163] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.090163] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.090746] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.090746] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.090746] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9c1fae21-16d8-4377-8f41-e9c9154ee27f, please check neutron logs for more information. [ 540.090746] env[63293]: ERROR nova.compute.manager [ 540.090746] env[63293]: Traceback (most recent call last): [ 540.090746] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.090746] env[63293]: listener.cb(fileno) [ 540.090746] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.090746] env[63293]: result = function(*args, **kwargs) [ 540.090746] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.090746] env[63293]: return func(*args, **kwargs) [ 540.090746] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.090746] env[63293]: raise e [ 540.090746] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.090746] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 540.090746] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.090746] env[63293]: created_port_ids = self._update_ports_for_instance( [ 540.090746] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.090746] env[63293]: with excutils.save_and_reraise_exception(): [ 540.090746] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.090746] env[63293]: self.force_reraise() [ 540.090746] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.090746] env[63293]: raise self.value [ 540.090746] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.090746] env[63293]: updated_port = self._update_port( [ 540.090746] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.090746] env[63293]: _ensure_no_port_binding_failure(port) [ 540.090746] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.090746] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.091514] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 9c1fae21-16d8-4377-8f41-e9c9154ee27f, please check neutron logs for more information. [ 540.091514] env[63293]: Removing descriptor: 17 [ 540.091514] env[63293]: ERROR nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9c1fae21-16d8-4377-8f41-e9c9154ee27f, please check neutron logs for more information. [ 540.091514] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] Traceback (most recent call last): [ 540.091514] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 540.091514] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] yield resources [ 540.091514] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.091514] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] self.driver.spawn(context, instance, image_meta, [ 540.091514] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 540.091514] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.091514] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.091514] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] vm_ref = self.build_virtual_machine(instance, [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] for vif in network_info: [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] return self._sync_wrapper(fn, *args, **kwargs) [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] self.wait() [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] self[:] = self._gt.wait() [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] return self._exit_event.wait() [ 540.091835] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] result = hub.switch() [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] return self.greenlet.switch() [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] result = function(*args, **kwargs) [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] return func(*args, **kwargs) [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] raise e [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] nwinfo = self.network_api.allocate_for_instance( [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.092204] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] created_port_ids = self._update_ports_for_instance( [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] with excutils.save_and_reraise_exception(): [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] self.force_reraise() [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] raise self.value [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] updated_port = self._update_port( [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] _ensure_no_port_binding_failure(port) [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.092547] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] raise exception.PortBindingFailed(port_id=port['id']) [ 540.092863] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] nova.exception.PortBindingFailed: Binding failed for port 9c1fae21-16d8-4377-8f41-e9c9154ee27f, please check neutron logs for more information. [ 540.092863] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] [ 540.092863] env[63293]: INFO nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Terminating instance [ 540.097316] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Acquiring lock "refresh_cache-dddc6041-4436-48da-8ed8-45be063b566b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.097480] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Acquired lock "refresh_cache-dddc6041-4436-48da-8ed8-45be063b566b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.097682] env[63293]: DEBUG nova.network.neutron [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 540.099402] env[63293]: DEBUG nova.scheduler.client.report [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 540.106110] env[63293]: DEBUG nova.virt.hardware [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 540.106110] env[63293]: DEBUG nova.virt.hardware [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 540.106110] env[63293]: DEBUG nova.virt.hardware [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 540.106283] env[63293]: DEBUG nova.virt.hardware [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 540.106283] env[63293]: DEBUG nova.virt.hardware [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 540.106283] env[63293]: DEBUG nova.virt.hardware [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 540.106283] env[63293]: DEBUG nova.virt.hardware [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 540.106283] env[63293]: DEBUG nova.virt.hardware [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 540.106438] env[63293]: DEBUG nova.virt.hardware [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 540.106438] env[63293]: DEBUG nova.virt.hardware [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 540.109380] env[63293]: DEBUG nova.virt.hardware [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 540.109380] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e15bdfed-5f48-405d-bc29-2fea884a017f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.117028] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddfa345-f727-446f-926a-ad92339fe2ac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.293410] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "510b9ea6-04f1-4704-b5e4-e19db57769a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.293828] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "510b9ea6-04f1-4704-b5e4-e19db57769a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.333633] env[63293]: DEBUG nova.compute.manager [req-50919ce0-8e27-4507-a060-51052e5f828e req-4c594ce0-d9ba-45be-acf4-9ec9cffaaa2e service nova] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Received event network-changed-005e83b3-4351-48e5-aefc-c6620e43ea0b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 540.333633] env[63293]: DEBUG nova.compute.manager [req-50919ce0-8e27-4507-a060-51052e5f828e req-4c594ce0-d9ba-45be-acf4-9ec9cffaaa2e service nova] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Refreshing instance network info cache due to event network-changed-005e83b3-4351-48e5-aefc-c6620e43ea0b. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 540.333835] env[63293]: DEBUG oslo_concurrency.lockutils [req-50919ce0-8e27-4507-a060-51052e5f828e req-4c594ce0-d9ba-45be-acf4-9ec9cffaaa2e service nova] Acquiring lock "refresh_cache-f255ca12-7110-4f73-a6ff-d6bcc122486b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.333975] env[63293]: DEBUG oslo_concurrency.lockutils [req-50919ce0-8e27-4507-a060-51052e5f828e req-4c594ce0-d9ba-45be-acf4-9ec9cffaaa2e service nova] Acquired lock "refresh_cache-f255ca12-7110-4f73-a6ff-d6bcc122486b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.334199] env[63293]: DEBUG nova.network.neutron [req-50919ce0-8e27-4507-a060-51052e5f828e req-4c594ce0-d9ba-45be-acf4-9ec9cffaaa2e service nova] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Refreshing network info cache for port 005e83b3-4351-48e5-aefc-c6620e43ea0b {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 540.433559] env[63293]: INFO nova.compute.manager [-] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Took 1.05 seconds to deallocate network for instance. [ 540.438019] env[63293]: DEBUG nova.compute.claims [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 540.438019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.581614] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327420, 'name': CreateVM_Task, 'duration_secs': 0.298703} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.581754] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 540.583500] env[63293]: DEBUG oslo_vmware.service [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ae983b-4a7f-4c61-9a7e-66f0d0c767ab {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.590667] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.590829] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.591214] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 540.591486] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d6f8926-4776-49c6-8127-4cd1eacec1f5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.595866] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 540.595866] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d73776-fdfe-379b-06a9-75c4bd4e1fb2" [ 540.595866] env[63293]: _type = "Task" [ 540.595866] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.604042] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d73776-fdfe-379b-06a9-75c4bd4e1fb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.611556] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.612037] env[63293]: DEBUG nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 540.616616] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.615s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.630867] env[63293]: DEBUG nova.network.neutron [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.777266] env[63293]: DEBUG nova.network.neutron [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.864384] env[63293]: DEBUG nova.network.neutron [req-50919ce0-8e27-4507-a060-51052e5f828e req-4c594ce0-d9ba-45be-acf4-9ec9cffaaa2e service nova] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.063582] env[63293]: DEBUG nova.network.neutron [req-50919ce0-8e27-4507-a060-51052e5f828e req-4c594ce0-d9ba-45be-acf4-9ec9cffaaa2e service nova] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.112324] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.112324] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 541.112324] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.112324] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.112578] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 541.112578] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59998088-df84-452e-bdf6-4ba44ac4f07b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.121593] env[63293]: DEBUG nova.compute.utils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 541.128180] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 541.128355] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 541.129116] env[63293]: DEBUG nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 541.129324] env[63293]: DEBUG nova.network.neutron [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 541.131523] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c04156f-53b3-42ee-8a87-a54d385177de {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.138831] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-912e6eff-905e-44d4-bc02-94cb21a61643 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.144784] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 541.144784] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52686b4f-3d23-78ae-5720-9e5cd9618614" [ 541.144784] env[63293]: _type = "Task" [ 541.144784] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.158093] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52686b4f-3d23-78ae-5720-9e5cd9618614, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.231724] env[63293]: DEBUG nova.policy [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3bb00827293c4b74a6f86f8db3a09cfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9169d8b45ed148b8935d1f8ca96ac3e0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 541.280124] env[63293]: ERROR nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f5b367dd-acd0-41f7-81bf-cfdcdafe0e82, please check neutron logs for more information. [ 541.280124] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 541.280124] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.280124] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 541.280124] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.280124] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 541.280124] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.280124] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 541.280124] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.280124] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 541.280124] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.280124] env[63293]: ERROR nova.compute.manager raise self.value [ 541.280124] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.280124] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 541.280124] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.280124] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 541.280601] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.280601] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 541.280601] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f5b367dd-acd0-41f7-81bf-cfdcdafe0e82, please check neutron logs for more information. [ 541.280601] env[63293]: ERROR nova.compute.manager [ 541.280601] env[63293]: Traceback (most recent call last): [ 541.280601] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 541.280601] env[63293]: listener.cb(fileno) [ 541.280601] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.280601] env[63293]: result = function(*args, **kwargs) [ 541.280601] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 541.280601] env[63293]: return func(*args, **kwargs) [ 541.280601] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.280601] env[63293]: raise e [ 541.280601] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.280601] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 541.280601] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.280601] env[63293]: created_port_ids = self._update_ports_for_instance( [ 541.280601] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.280601] env[63293]: with excutils.save_and_reraise_exception(): [ 541.280601] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.280601] env[63293]: self.force_reraise() [ 541.280601] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.280601] env[63293]: raise self.value [ 541.280601] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.280601] env[63293]: updated_port = self._update_port( [ 541.280601] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.280601] env[63293]: _ensure_no_port_binding_failure(port) [ 541.280601] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.280601] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 541.281360] env[63293]: nova.exception.PortBindingFailed: Binding failed for port f5b367dd-acd0-41f7-81bf-cfdcdafe0e82, please check neutron logs for more information. [ 541.281360] env[63293]: Removing descriptor: 18 [ 541.281360] env[63293]: ERROR nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f5b367dd-acd0-41f7-81bf-cfdcdafe0e82, please check neutron logs for more information. [ 541.281360] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Traceback (most recent call last): [ 541.281360] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 541.281360] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] yield resources [ 541.281360] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.281360] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] self.driver.spawn(context, instance, image_meta, [ 541.281360] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 541.281360] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.281360] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.281360] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] vm_ref = self.build_virtual_machine(instance, [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] for vif in network_info: [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] return self._sync_wrapper(fn, *args, **kwargs) [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] self.wait() [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] self[:] = self._gt.wait() [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] return self._exit_event.wait() [ 541.281677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] result = hub.switch() [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] return self.greenlet.switch() [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] result = function(*args, **kwargs) [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] return func(*args, **kwargs) [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] raise e [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] nwinfo = self.network_api.allocate_for_instance( [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.281990] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] created_port_ids = self._update_ports_for_instance( [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] with excutils.save_and_reraise_exception(): [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] self.force_reraise() [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] raise self.value [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] updated_port = self._update_port( [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] _ensure_no_port_binding_failure(port) [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.282395] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] raise exception.PortBindingFailed(port_id=port['id']) [ 541.282720] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] nova.exception.PortBindingFailed: Binding failed for port f5b367dd-acd0-41f7-81bf-cfdcdafe0e82, please check neutron logs for more information. [ 541.282720] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] [ 541.282720] env[63293]: INFO nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Terminating instance [ 541.283301] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Releasing lock "refresh_cache-dddc6041-4436-48da-8ed8-45be063b566b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.283665] env[63293]: DEBUG nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 541.283848] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 541.284389] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquiring lock "refresh_cache-9982e59e-65af-44c7-bdd9-056565107a76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.284709] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquired lock "refresh_cache-9982e59e-65af-44c7-bdd9-056565107a76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.284709] env[63293]: DEBUG nova.network.neutron [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 541.286379] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-25396eae-492d-493f-b159-05a47608b6e0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.300902] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ce70d0-e7ef-4507-8e76-b373df29d953 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.332807] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dddc6041-4436-48da-8ed8-45be063b566b could not be found. [ 541.332807] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 541.332807] env[63293]: INFO nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 541.332807] env[63293]: DEBUG oslo.service.loopingcall [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.335205] env[63293]: DEBUG nova.compute.manager [-] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.335205] env[63293]: DEBUG nova.network.neutron [-] [instance: dddc6041-4436-48da-8ed8-45be063b566b] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 541.367499] env[63293]: DEBUG nova.network.neutron [-] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.568542] env[63293]: DEBUG oslo_concurrency.lockutils [req-50919ce0-8e27-4507-a060-51052e5f828e req-4c594ce0-d9ba-45be-acf4-9ec9cffaaa2e service nova] Releasing lock "refresh_cache-f255ca12-7110-4f73-a6ff-d6bcc122486b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.570372] env[63293]: DEBUG nova.compute.manager [req-50919ce0-8e27-4507-a060-51052e5f828e req-4c594ce0-d9ba-45be-acf4-9ec9cffaaa2e service nova] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Received event network-vif-deleted-005e83b3-4351-48e5-aefc-c6620e43ea0b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 541.576562] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee057536-597b-4e96-bee6-5a007935fa2a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.584754] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d77840a-4288-423f-9f57-65169ef5cc26 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.619106] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81acbcee-9617-4e88-869f-2cb90f308088 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.627374] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca0f450-df8e-4ca0-a558-2cdb3b8f7b0a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.631717] env[63293]: DEBUG nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 541.645272] env[63293]: DEBUG nova.compute.provider_tree [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.655047] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Preparing fetch location {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 541.655120] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Creating directory with path [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 541.656205] env[63293]: DEBUG nova.network.neutron [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Successfully created port: 2b781a20-f80f-43a9-aebd-2fe700fc0ae4 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 541.658593] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5779d92-c1d5-48a5-9a67-c03e430b31d7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.672278] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Created directory with path [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 541.672646] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Fetch image to [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 541.672646] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Downloading image file data 1427aa66-93c7-49c6-9e28-dc7fe851dced to [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk on the data store datastore1 {{(pid=63293) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 541.674334] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7468b0-4049-45a1-9e6c-a73c5df49e3a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.681549] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d1df33-4bf6-4311-b457-11c4696c09c1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.691316] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fa3699-b7b4-4236-add1-b7aa81658e91 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.726552] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de9b062-e86f-45e4-a1b2-c790c77fa210 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.732098] env[63293]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-bed5ed62-c4c5-45c8-ad83-9c4ca695104c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.820828] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Downloading image file data 1427aa66-93c7-49c6-9e28-dc7fe851dced to the data store datastore1 {{(pid=63293) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 541.869803] env[63293]: DEBUG nova.network.neutron [-] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.892495] env[63293]: DEBUG oslo_vmware.rw_handles [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63293) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 541.953390] env[63293]: DEBUG nova.network.neutron [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.151058] env[63293]: DEBUG nova.scheduler.client.report [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 542.176031] env[63293]: DEBUG nova.network.neutron [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.373387] env[63293]: INFO nova.compute.manager [-] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Took 1.04 seconds to deallocate network for instance. [ 542.376482] env[63293]: DEBUG nova.compute.claims [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 542.376822] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.572324] env[63293]: DEBUG oslo_vmware.rw_handles [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Completed reading data from the image iterator. {{(pid=63293) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 542.572324] env[63293]: DEBUG oslo_vmware.rw_handles [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 542.642061] env[63293]: DEBUG nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 542.664268] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.047s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.665106] env[63293]: ERROR nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6a6d81e5-3c63-425b-938b-2294443c2fd4, please check neutron logs for more information. [ 542.665106] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Traceback (most recent call last): [ 542.665106] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 542.665106] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] self.driver.spawn(context, instance, image_meta, [ 542.665106] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 542.665106] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] self._vmops.spawn(context, instance, image_meta, injected_files, [ 542.665106] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 542.665106] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] vm_ref = self.build_virtual_machine(instance, [ 542.665106] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 542.665106] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] vif_infos = vmwarevif.get_vif_info(self._session, [ 542.665106] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] for vif in network_info: [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] return self._sync_wrapper(fn, *args, **kwargs) [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] self.wait() [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] self[:] = self._gt.wait() [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] return self._exit_event.wait() [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] result = hub.switch() [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 542.665943] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] return self.greenlet.switch() [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] result = function(*args, **kwargs) [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] return func(*args, **kwargs) [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] raise e [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] nwinfo = self.network_api.allocate_for_instance( [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] created_port_ids = self._update_ports_for_instance( [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] with excutils.save_and_reraise_exception(): [ 542.667660] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.668020] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] self.force_reraise() [ 542.668020] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.668020] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] raise self.value [ 542.668020] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 542.668020] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] updated_port = self._update_port( [ 542.668020] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.668020] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] _ensure_no_port_binding_failure(port) [ 542.668020] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.668020] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] raise exception.PortBindingFailed(port_id=port['id']) [ 542.668020] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] nova.exception.PortBindingFailed: Binding failed for port 6a6d81e5-3c63-425b-938b-2294443c2fd4, please check neutron logs for more information. [ 542.668020] env[63293]: ERROR nova.compute.manager [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] [ 542.668310] env[63293]: DEBUG nova.compute.utils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Binding failed for port 6a6d81e5-3c63-425b-938b-2294443c2fd4, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 542.668310] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.762s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.668310] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.668310] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63293) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 542.668310] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.911s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.675712] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6b0936-f2aa-402f-885b-93109160efe4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.684134] env[63293]: DEBUG nova.virt.hardware [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 542.684364] env[63293]: DEBUG nova.virt.hardware [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 542.684534] env[63293]: DEBUG nova.virt.hardware [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 542.684714] env[63293]: DEBUG nova.virt.hardware [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 542.684855] env[63293]: DEBUG nova.virt.hardware [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 542.685147] env[63293]: DEBUG nova.virt.hardware [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 542.685238] env[63293]: DEBUG nova.virt.hardware [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 542.685396] env[63293]: DEBUG nova.virt.hardware [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 542.685578] env[63293]: DEBUG nova.virt.hardware [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 542.686925] env[63293]: DEBUG nova.virt.hardware [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 542.686925] env[63293]: DEBUG nova.virt.hardware [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 542.686925] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Releasing lock "refresh_cache-9982e59e-65af-44c7-bdd9-056565107a76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.686925] env[63293]: DEBUG nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 542.686925] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 542.687755] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72bd7b49-52d7-41d5-aa04-1dd7404f2920 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.693752] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cbee5788-6c35-400e-baa8-1edf67075b1a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.700419] env[63293]: DEBUG nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Build of instance d9d24a3a-078a-49a6-a02c-3fb0e09ef772 was re-scheduled: Binding failed for port 6a6d81e5-3c63-425b-938b-2294443c2fd4, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 542.700580] env[63293]: DEBUG nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 542.700940] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquiring lock "refresh_cache-d9d24a3a-078a-49a6-a02c-3fb0e09ef772" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.700940] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquired lock "refresh_cache-d9d24a3a-078a-49a6-a02c-3fb0e09ef772" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.701424] env[63293]: DEBUG nova.network.neutron [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 542.715724] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0425cf3-69b8-4a4f-9e9e-4c4746b44b3b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.722653] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Downloaded image file data 1427aa66-93c7-49c6-9e28-dc7fe851dced to vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk on the data store datastore1 {{(pid=63293) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 542.725887] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Caching image {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 542.726151] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Copying Virtual Disk [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk to [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 542.727425] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f8f072-7445-488b-a69b-e6cca3ffabe9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.732805] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-71fbe21f-1ecb-4003-95f8-7ec01272f5b0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.758782] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524f6e48-3e37-42bb-be51-05041e9496b0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.765013] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fe90fe-29a6-4b6f-ab11-ce4a1d219246 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.778514] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 542.778514] env[63293]: value = "task-1327421" [ 542.778514] env[63293]: _type = "Task" [ 542.778514] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.790147] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224a6f70-05a4-4386-b4f4-6606a31f7add {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.794198] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9982e59e-65af-44c7-bdd9-056565107a76 could not be found. [ 542.794569] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 542.794697] env[63293]: INFO nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Took 0.11 seconds to destroy the instance on the hypervisor. [ 542.794943] env[63293]: DEBUG oslo.service.loopingcall [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.795975] env[63293]: DEBUG nova.compute.manager [-] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 542.796106] env[63293]: DEBUG nova.network.neutron [-] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 542.801769] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327421, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.829382] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181535MB free_disk=132GB free_vcpus=48 pci_devices=None {{(pid=63293) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 542.829549] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.851594] env[63293]: DEBUG nova.network.neutron [-] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.962876] env[63293]: DEBUG nova.compute.manager [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Received event network-changed-9c1fae21-16d8-4377-8f41-e9c9154ee27f {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.964310] env[63293]: DEBUG nova.compute.manager [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Refreshing instance network info cache due to event network-changed-9c1fae21-16d8-4377-8f41-e9c9154ee27f. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 542.965196] env[63293]: DEBUG oslo_concurrency.lockutils [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] Acquiring lock "refresh_cache-dddc6041-4436-48da-8ed8-45be063b566b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.965196] env[63293]: DEBUG oslo_concurrency.lockutils [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] Acquired lock "refresh_cache-dddc6041-4436-48da-8ed8-45be063b566b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.965196] env[63293]: DEBUG nova.network.neutron [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Refreshing network info cache for port 9c1fae21-16d8-4377-8f41-e9c9154ee27f {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 543.245329] env[63293]: DEBUG nova.network.neutron [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.290570] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327421, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.356115] env[63293]: DEBUG nova.network.neutron [-] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.456529] env[63293]: DEBUG nova.network.neutron [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.510297] env[63293]: DEBUG nova.network.neutron [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.617306] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df004096-6730-4077-964e-7fecb70610de {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.627448] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabcec4e-e8b5-4dd8-8ea6-5974553da9b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.668083] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd62c5ca-3fb5-4710-808c-d279ec2b8c73 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.676353] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35aa4c48-6b7e-4802-8880-d2d79d974dd7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.691536] env[63293]: DEBUG nova.compute.provider_tree [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.792437] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327421, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.838622] env[63293]: DEBUG nova.network.neutron [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.862595] env[63293]: INFO nova.compute.manager [-] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Took 1.07 seconds to deallocate network for instance. [ 543.866050] env[63293]: DEBUG nova.compute.claims [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 543.866279] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.962707] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Releasing lock "refresh_cache-d9d24a3a-078a-49a6-a02c-3fb0e09ef772" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.963028] env[63293]: DEBUG nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 543.963291] env[63293]: DEBUG nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 543.963358] env[63293]: DEBUG nova.network.neutron [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.021119] env[63293]: DEBUG nova.network.neutron [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.196590] env[63293]: DEBUG nova.scheduler.client.report [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 544.295838] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327421, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.344225] env[63293]: DEBUG oslo_concurrency.lockutils [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] Releasing lock "refresh_cache-dddc6041-4436-48da-8ed8-45be063b566b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.344225] env[63293]: DEBUG nova.compute.manager [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Received event network-vif-deleted-9c1fae21-16d8-4377-8f41-e9c9154ee27f {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.344225] env[63293]: DEBUG nova.compute.manager [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Received event network-changed-f5b367dd-acd0-41f7-81bf-cfdcdafe0e82 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.344225] env[63293]: DEBUG nova.compute.manager [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Refreshing instance network info cache due to event network-changed-f5b367dd-acd0-41f7-81bf-cfdcdafe0e82. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 544.344225] env[63293]: DEBUG oslo_concurrency.lockutils [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] Acquiring lock "refresh_cache-9982e59e-65af-44c7-bdd9-056565107a76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.344541] env[63293]: DEBUG oslo_concurrency.lockutils [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] Acquired lock "refresh_cache-9982e59e-65af-44c7-bdd9-056565107a76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.344541] env[63293]: DEBUG nova.network.neutron [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Refreshing network info cache for port f5b367dd-acd0-41f7-81bf-cfdcdafe0e82 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 544.432697] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "eaaf2515-cc4b-42c7-aacd-14074c760924" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.432943] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "eaaf2515-cc4b-42c7-aacd-14074c760924" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.529284] env[63293]: DEBUG nova.network.neutron [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.706515] env[63293]: ERROR nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2b781a20-f80f-43a9-aebd-2fe700fc0ae4, please check neutron logs for more information. [ 544.706515] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 544.706515] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.706515] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 544.706515] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.706515] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 544.706515] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.706515] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 544.706515] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.706515] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 544.706515] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.706515] env[63293]: ERROR nova.compute.manager raise self.value [ 544.706515] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.706515] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 544.706515] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.706515] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 544.709077] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.709077] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 544.709077] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2b781a20-f80f-43a9-aebd-2fe700fc0ae4, please check neutron logs for more information. [ 544.709077] env[63293]: ERROR nova.compute.manager [ 544.709077] env[63293]: Traceback (most recent call last): [ 544.709077] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 544.709077] env[63293]: listener.cb(fileno) [ 544.709077] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.709077] env[63293]: result = function(*args, **kwargs) [ 544.709077] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 544.709077] env[63293]: return func(*args, **kwargs) [ 544.709077] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.709077] env[63293]: raise e [ 544.709077] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.709077] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 544.709077] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.709077] env[63293]: created_port_ids = self._update_ports_for_instance( [ 544.709077] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.709077] env[63293]: with excutils.save_and_reraise_exception(): [ 544.709077] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.709077] env[63293]: self.force_reraise() [ 544.709077] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.709077] env[63293]: raise self.value [ 544.709077] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.709077] env[63293]: updated_port = self._update_port( [ 544.709077] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.709077] env[63293]: _ensure_no_port_binding_failure(port) [ 544.709077] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.709077] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 544.710170] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 2b781a20-f80f-43a9-aebd-2fe700fc0ae4, please check neutron logs for more information. [ 544.710170] env[63293]: Removing descriptor: 17 [ 544.710170] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.039s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.710170] env[63293]: ERROR nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ead43854-09d6-4e20-ab13-c5e48dd12379, please check neutron logs for more information. [ 544.710170] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Traceback (most recent call last): [ 544.710170] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.710170] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] self.driver.spawn(context, instance, image_meta, [ 544.710170] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 544.710170] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.710170] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] vm_ref = self.build_virtual_machine(instance, [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] for vif in network_info: [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] return self._sync_wrapper(fn, *args, **kwargs) [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] self.wait() [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] self[:] = self._gt.wait() [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.710436] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] return self._exit_event.wait() [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] result = hub.switch() [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] return self.greenlet.switch() [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] result = function(*args, **kwargs) [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] return func(*args, **kwargs) [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] raise e [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] nwinfo = self.network_api.allocate_for_instance( [ 544.710767] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] created_port_ids = self._update_ports_for_instance( [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] with excutils.save_and_reraise_exception(): [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] self.force_reraise() [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] raise self.value [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] updated_port = self._update_port( [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] _ensure_no_port_binding_failure(port) [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.711096] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] raise exception.PortBindingFailed(port_id=port['id']) [ 544.711954] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] nova.exception.PortBindingFailed: Binding failed for port ead43854-09d6-4e20-ab13-c5e48dd12379, please check neutron logs for more information. [ 544.711954] env[63293]: ERROR nova.compute.manager [instance: f9e64134-f233-48a3-afc3-be7765a67c35] [ 544.711954] env[63293]: DEBUG nova.compute.utils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Binding failed for port ead43854-09d6-4e20-ab13-c5e48dd12379, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 544.711954] env[63293]: ERROR nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2b781a20-f80f-43a9-aebd-2fe700fc0ae4, please check neutron logs for more information. [ 544.711954] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] Traceback (most recent call last): [ 544.711954] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 544.711954] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] yield resources [ 544.711954] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.711954] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] self.driver.spawn(context, instance, image_meta, [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] vm_ref = self.build_virtual_machine(instance, [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] for vif in network_info: [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] return self._sync_wrapper(fn, *args, **kwargs) [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] self.wait() [ 544.712227] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] self[:] = self._gt.wait() [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] return self._exit_event.wait() [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] result = hub.switch() [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] return self.greenlet.switch() [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] result = function(*args, **kwargs) [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] return func(*args, **kwargs) [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.712543] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] raise e [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] nwinfo = self.network_api.allocate_for_instance( [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] created_port_ids = self._update_ports_for_instance( [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] with excutils.save_and_reraise_exception(): [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] self.force_reraise() [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] raise self.value [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] updated_port = self._update_port( [ 544.712849] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.713178] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] _ensure_no_port_binding_failure(port) [ 544.713178] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.713178] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] raise exception.PortBindingFailed(port_id=port['id']) [ 544.713178] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] nova.exception.PortBindingFailed: Binding failed for port 2b781a20-f80f-43a9-aebd-2fe700fc0ae4, please check neutron logs for more information. [ 544.713178] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] [ 544.713178] env[63293]: INFO nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Terminating instance [ 544.713178] env[63293]: DEBUG nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Build of instance f9e64134-f233-48a3-afc3-be7765a67c35 was re-scheduled: Binding failed for port ead43854-09d6-4e20-ab13-c5e48dd12379, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 544.713178] env[63293]: DEBUG nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 544.713443] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Acquiring lock "refresh_cache-f9e64134-f233-48a3-afc3-be7765a67c35" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.713600] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Acquired lock "refresh_cache-f9e64134-f233-48a3-afc3-be7765a67c35" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.713758] env[63293]: DEBUG nova.network.neutron [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 544.715230] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.588s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.730601] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "refresh_cache-dccd4149-a888-4910-95c8-8d0028338dee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.730783] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquired lock "refresh_cache-dccd4149-a888-4910-95c8-8d0028338dee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.730957] env[63293]: DEBUG nova.network.neutron [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 544.797555] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327421, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.688967} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.797800] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Copied Virtual Disk [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk to [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 544.798014] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleting the datastore file [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced/tmp-sparse.vmdk {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 544.798281] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1fc1ff7-a690-4374-8eee-ae36e684ddd0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.806385] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 544.806385] env[63293]: value = "task-1327422" [ 544.806385] env[63293]: _type = "Task" [ 544.806385] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.815333] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327422, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.868383] env[63293]: DEBUG nova.network.neutron [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.978518] env[63293]: DEBUG nova.network.neutron [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.031097] env[63293]: INFO nova.compute.manager [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: d9d24a3a-078a-49a6-a02c-3fb0e09ef772] Took 1.07 seconds to deallocate network for instance. [ 545.266526] env[63293]: DEBUG nova.network.neutron [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.277222] env[63293]: DEBUG nova.network.neutron [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.325841] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327422, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023624} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.325841] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 545.325841] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Moving file from [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017/1427aa66-93c7-49c6-9e28-dc7fe851dced to [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced. {{(pid=63293) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 545.326182] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-3c66bb9e-9d2f-488d-bc69-2061258bfc08 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.337991] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 545.337991] env[63293]: value = "task-1327423" [ 545.337991] env[63293]: _type = "Task" [ 545.337991] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.351548] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327423, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.367035] env[63293]: DEBUG nova.network.neutron [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.409347] env[63293]: DEBUG nova.network.neutron [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.482262] env[63293]: DEBUG oslo_concurrency.lockutils [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] Releasing lock "refresh_cache-9982e59e-65af-44c7-bdd9-056565107a76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.482591] env[63293]: DEBUG nova.compute.manager [req-991ee390-1873-4b3a-a7c8-7c93662ba95c req-5956e544-4f31-4acf-afe1-e9cb3d635acb service nova] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Received event network-vif-deleted-f5b367dd-acd0-41f7-81bf-cfdcdafe0e82 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.620494] env[63293]: DEBUG nova.compute.manager [req-8c01ad61-954d-401b-84e2-0921f944ea1d req-6d09113b-1d0e-4b15-bb1d-bac9d41357ba service nova] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Received event network-changed-2b781a20-f80f-43a9-aebd-2fe700fc0ae4 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.620494] env[63293]: DEBUG nova.compute.manager [req-8c01ad61-954d-401b-84e2-0921f944ea1d req-6d09113b-1d0e-4b15-bb1d-bac9d41357ba service nova] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Refreshing instance network info cache due to event network-changed-2b781a20-f80f-43a9-aebd-2fe700fc0ae4. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 545.621136] env[63293]: DEBUG oslo_concurrency.lockutils [req-8c01ad61-954d-401b-84e2-0921f944ea1d req-6d09113b-1d0e-4b15-bb1d-bac9d41357ba service nova] Acquiring lock "refresh_cache-dccd4149-a888-4910-95c8-8d0028338dee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.673023] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07670e76-bada-4ece-ad7e-4c172f0fa177 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.683852] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09385be5-407a-43a9-b660-62c2b5d6d11d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.720904] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f338151c-d0b3-4059-a276-56dd2895166a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.729038] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a821f7-d1a7-4faf-a24a-e35a032464e3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.744102] env[63293]: DEBUG nova.compute.provider_tree [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.849873] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327423, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.028959} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.850458] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] File moved {{(pid=63293) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 545.850598] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Cleaning up location [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017 {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 545.850764] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleting the datastore file [datastore1] vmware_temp/25b79158-3d95-4173-bdb4-468f71aca017 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 545.852225] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63a23ace-a730-48e5-9e53-44f8e633b38a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.858381] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 545.858381] env[63293]: value = "task-1327424" [ 545.858381] env[63293]: _type = "Task" [ 545.858381] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.867434] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327424, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.875324] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Releasing lock "refresh_cache-f9e64134-f233-48a3-afc3-be7765a67c35" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.875324] env[63293]: DEBUG nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 545.875324] env[63293]: DEBUG nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.875324] env[63293]: DEBUG nova.network.neutron [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 545.898183] env[63293]: DEBUG nova.network.neutron [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.910503] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Releasing lock "refresh_cache-dccd4149-a888-4910-95c8-8d0028338dee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.910940] env[63293]: DEBUG nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 545.911140] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 545.911695] env[63293]: DEBUG oslo_concurrency.lockutils [req-8c01ad61-954d-401b-84e2-0921f944ea1d req-6d09113b-1d0e-4b15-bb1d-bac9d41357ba service nova] Acquired lock "refresh_cache-dccd4149-a888-4910-95c8-8d0028338dee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.911924] env[63293]: DEBUG nova.network.neutron [req-8c01ad61-954d-401b-84e2-0921f944ea1d req-6d09113b-1d0e-4b15-bb1d-bac9d41357ba service nova] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Refreshing network info cache for port 2b781a20-f80f-43a9-aebd-2fe700fc0ae4 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 545.912823] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ccd1c16-25e0-4c81-85ed-cd8d1daab21d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.922262] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b7daa9-2dab-4def-a93c-ba2edd557036 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.948337] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dccd4149-a888-4910-95c8-8d0028338dee could not be found. [ 545.948579] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 545.948759] env[63293]: INFO nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Took 0.04 seconds to destroy the instance on the hypervisor. [ 545.949210] env[63293]: DEBUG oslo.service.loopingcall [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 545.949293] env[63293]: DEBUG nova.compute.manager [-] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.949330] env[63293]: DEBUG nova.network.neutron [-] [instance: dccd4149-a888-4910-95c8-8d0028338dee] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 545.977187] env[63293]: DEBUG nova.network.neutron [-] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.083692] env[63293]: INFO nova.scheduler.client.report [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Deleted allocations for instance d9d24a3a-078a-49a6-a02c-3fb0e09ef772 [ 546.247297] env[63293]: DEBUG nova.scheduler.client.report [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.377890] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327424, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025456} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.377890] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 546.377890] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c1b8753-e4bc-4e8a-88b7-e4538394ae0c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.382851] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 546.382851] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522ccdc0-e87f-b3d5-65b2-d63b753e240e" [ 546.382851] env[63293]: _type = "Task" [ 546.382851] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.391671] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522ccdc0-e87f-b3d5-65b2-d63b753e240e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.405017] env[63293]: DEBUG nova.network.neutron [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.447625] env[63293]: DEBUG nova.network.neutron [req-8c01ad61-954d-401b-84e2-0921f944ea1d req-6d09113b-1d0e-4b15-bb1d-bac9d41357ba service nova] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.482697] env[63293]: DEBUG nova.network.neutron [-] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.600103] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fb54f72-3346-44ca-9307-f6adb54b0204 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "d9d24a3a-078a-49a6-a02c-3fb0e09ef772" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.085s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.655297] env[63293]: DEBUG nova.network.neutron [req-8c01ad61-954d-401b-84e2-0921f944ea1d req-6d09113b-1d0e-4b15-bb1d-bac9d41357ba service nova] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.756022] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.038s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.756022] env[63293]: ERROR nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9f614a5-9c36-44bc-8b23-6b8cc5e475d9, please check neutron logs for more information. [ 546.756022] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Traceback (most recent call last): [ 546.756022] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 546.756022] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] self.driver.spawn(context, instance, image_meta, [ 546.756022] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 546.756022] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 546.756022] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 546.756022] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] vm_ref = self.build_virtual_machine(instance, [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] for vif in network_info: [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] return self._sync_wrapper(fn, *args, **kwargs) [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] self.wait() [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] self[:] = self._gt.wait() [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] return self._exit_event.wait() [ 546.756808] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] result = hub.switch() [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] return self.greenlet.switch() [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] result = function(*args, **kwargs) [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] return func(*args, **kwargs) [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] raise e [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] nwinfo = self.network_api.allocate_for_instance( [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.757178] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] created_port_ids = self._update_ports_for_instance( [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] with excutils.save_and_reraise_exception(): [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] self.force_reraise() [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] raise self.value [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] updated_port = self._update_port( [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] _ensure_no_port_binding_failure(port) [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.757538] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] raise exception.PortBindingFailed(port_id=port['id']) [ 546.757853] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] nova.exception.PortBindingFailed: Binding failed for port a9f614a5-9c36-44bc-8b23-6b8cc5e475d9, please check neutron logs for more information. [ 546.757853] env[63293]: ERROR nova.compute.manager [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] [ 546.757853] env[63293]: DEBUG nova.compute.utils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Binding failed for port a9f614a5-9c36-44bc-8b23-6b8cc5e475d9, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 546.757853] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.648s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.764797] env[63293]: DEBUG nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Build of instance c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4 was re-scheduled: Binding failed for port a9f614a5-9c36-44bc-8b23-6b8cc5e475d9, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 546.765287] env[63293]: DEBUG nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 546.765529] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Acquiring lock "refresh_cache-c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.765703] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Acquired lock "refresh_cache-c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.765877] env[63293]: DEBUG nova.network.neutron [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 546.896102] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522ccdc0-e87f-b3d5-65b2-d63b753e240e, 'name': SearchDatastore_Task, 'duration_secs': 0.008764} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.896102] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.896102] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 546.896262] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49cfd1b0-8b9b-47c2-a873-d7619b3b6f3b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.907935] env[63293]: INFO nova.compute.manager [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] [instance: f9e64134-f233-48a3-afc3-be7765a67c35] Took 1.04 seconds to deallocate network for instance. [ 546.910640] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 546.910640] env[63293]: value = "task-1327425" [ 546.910640] env[63293]: _type = "Task" [ 546.910640] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.923193] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327425, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.986125] env[63293]: INFO nova.compute.manager [-] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Took 1.04 seconds to deallocate network for instance. [ 546.989402] env[63293]: DEBUG nova.compute.claims [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 546.990366] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.105505] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 547.158972] env[63293]: DEBUG oslo_concurrency.lockutils [req-8c01ad61-954d-401b-84e2-0921f944ea1d req-6d09113b-1d0e-4b15-bb1d-bac9d41357ba service nova] Releasing lock "refresh_cache-dccd4149-a888-4910-95c8-8d0028338dee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.158972] env[63293]: DEBUG nova.compute.manager [req-8c01ad61-954d-401b-84e2-0921f944ea1d req-6d09113b-1d0e-4b15-bb1d-bac9d41357ba service nova] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Received event network-vif-deleted-2b781a20-f80f-43a9-aebd-2fe700fc0ae4 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.228249] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "1814ebc5-7869-4ec6-8cf6-790daed5e271" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.228483] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "1814ebc5-7869-4ec6-8cf6-790daed5e271" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.330029] env[63293]: DEBUG nova.network.neutron [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.425191] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327425, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.525964] env[63293]: DEBUG nova.network.neutron [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.644260] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.718530] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85bcf3d4-3694-4c01-b4db-a9de27d98008 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.728633] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247e0580-4909-4245-addd-8e14890e46bd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.760459] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cfaad65-8ca4-47f1-aa4c-37ccb1c9bb5c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.768255] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611a4620-abba-4d6f-bf3e-ac8c4b0d4b48 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.782976] env[63293]: DEBUG nova.compute.provider_tree [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.929311] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327425, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.955945] env[63293]: INFO nova.scheduler.client.report [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Deleted allocations for instance f9e64134-f233-48a3-afc3-be7765a67c35 [ 548.028815] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Releasing lock "refresh_cache-c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.028815] env[63293]: DEBUG nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 548.029159] env[63293]: DEBUG nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.029159] env[63293]: DEBUG nova.network.neutron [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.079201] env[63293]: DEBUG nova.network.neutron [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.287750] env[63293]: DEBUG nova.scheduler.client.report [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.430280] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327425, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.518066} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.431084] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 548.431603] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 548.432025] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d97e017-6a11-4b30-ac0d-19da50b6eed0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.443391] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 548.443391] env[63293]: value = "task-1327426" [ 548.443391] env[63293]: _type = "Task" [ 548.443391] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.457380] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327426, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.467705] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c7a855d-4fe0-49a3-bcce-d16cf8902837 tempest-ServerDiagnosticsNegativeTest-196076807 tempest-ServerDiagnosticsNegativeTest-196076807-project-member] Lock "f9e64134-f233-48a3-afc3-be7765a67c35" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.288s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.585532] env[63293]: DEBUG nova.network.neutron [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.794873] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.037s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.794873] env[63293]: ERROR nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2251e186-6de4-4951-9bda-842fc18ba5bf, please check neutron logs for more information. [ 548.794873] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] Traceback (most recent call last): [ 548.794873] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.794873] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] self.driver.spawn(context, instance, image_meta, [ 548.794873] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 548.794873] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.794873] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.794873] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] vm_ref = self.build_virtual_machine(instance, [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] for vif in network_info: [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] return self._sync_wrapper(fn, *args, **kwargs) [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] self.wait() [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] self[:] = self._gt.wait() [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] return self._exit_event.wait() [ 548.795178] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] result = hub.switch() [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] return self.greenlet.switch() [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] result = function(*args, **kwargs) [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] return func(*args, **kwargs) [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] raise e [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] nwinfo = self.network_api.allocate_for_instance( [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 548.795658] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] created_port_ids = self._update_ports_for_instance( [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] with excutils.save_and_reraise_exception(): [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] self.force_reraise() [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] raise self.value [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] updated_port = self._update_port( [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] _ensure_no_port_binding_failure(port) [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.796012] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] raise exception.PortBindingFailed(port_id=port['id']) [ 548.796441] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] nova.exception.PortBindingFailed: Binding failed for port 2251e186-6de4-4951-9bda-842fc18ba5bf, please check neutron logs for more information. [ 548.796441] env[63293]: ERROR nova.compute.manager [instance: aa4611e2-6254-4983-8565-899aea616726] [ 548.796441] env[63293]: DEBUG nova.compute.utils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Binding failed for port 2251e186-6de4-4951-9bda-842fc18ba5bf, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 548.800351] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.591s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.803706] env[63293]: INFO nova.compute.claims [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 548.804916] env[63293]: DEBUG nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Build of instance aa4611e2-6254-4983-8565-899aea616726 was re-scheduled: Binding failed for port 2251e186-6de4-4951-9bda-842fc18ba5bf, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 548.805446] env[63293]: DEBUG nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 548.805720] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Acquiring lock "refresh_cache-aa4611e2-6254-4983-8565-899aea616726" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.805861] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Acquired lock "refresh_cache-aa4611e2-6254-4983-8565-899aea616726" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.806095] env[63293]: DEBUG nova.network.neutron [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 548.954331] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327426, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068365} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.954331] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 548.955466] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16ec79c-4ebb-462e-bcb5-9417cbebad3b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.980119] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 548.980556] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.985462] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2591daec-eb96-433e-b108-30677c1fba91 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.017142] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 549.017142] env[63293]: value = "task-1327427" [ 549.017142] env[63293]: _type = "Task" [ 549.017142] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.028241] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327427, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.089427] env[63293]: INFO nova.compute.manager [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] [instance: c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4] Took 1.06 seconds to deallocate network for instance. [ 549.144467] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Acquiring lock "63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.144688] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Lock "63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.340516] env[63293]: DEBUG nova.network.neutron [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.533038] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327427, 'name': ReconfigVM_Task, 'duration_secs': 0.275658} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.533038] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Reconfigured VM instance instance-00000002 to attach disk [datastore1] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 549.533377] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cbc43ee6-bbdb-4ee5-9c20-589aafd531f6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.540228] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.542189] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 549.542189] env[63293]: value = "task-1327428" [ 549.542189] env[63293]: _type = "Task" [ 549.542189] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.551841] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327428, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.659214] env[63293]: DEBUG nova.network.neutron [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.010155] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquiring lock "87c7bb2c-46ff-45f7-8967-14192f2b42e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.010357] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "87c7bb2c-46ff-45f7-8967-14192f2b42e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.063019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquiring lock "769c2643-57eb-456c-8083-997a8c9c48cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.063019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "769c2643-57eb-456c-8083-997a8c9c48cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.063463] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327428, 'name': Rename_Task, 'duration_secs': 0.153067} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.063709] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 550.063985] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b2a96cc-36b7-4366-93f3-4684a8121ffb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.070478] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 550.070478] env[63293]: value = "task-1327429" [ 550.070478] env[63293]: _type = "Task" [ 550.070478] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.082456] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327429, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.147293] env[63293]: INFO nova.scheduler.client.report [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Deleted allocations for instance c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4 [ 550.164472] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Releasing lock "refresh_cache-aa4611e2-6254-4983-8565-899aea616726" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.164784] env[63293]: DEBUG nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 550.164980] env[63293]: DEBUG nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 550.165203] env[63293]: DEBUG nova.network.neutron [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 550.199952] env[63293]: DEBUG nova.network.neutron [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.395192] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebafd754-342c-4d1c-baef-44c21f6745c1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.410332] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-072e0492-9eed-414e-8767-1a293432bee6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.453484] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5206d4f-4a51-45c5-8554-e78fcc9ed83a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.463243] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c455a6-6ec4-47fd-9688-f29018676fee {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.487104] env[63293]: DEBUG nova.compute.provider_tree [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.582529] env[63293]: DEBUG oslo_vmware.api [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327429, 'name': PowerOnVM_Task, 'duration_secs': 0.432642} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.582529] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 550.582890] env[63293]: DEBUG nova.compute.manager [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 550.584027] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4183d37a-292f-4142-9619-a0167ce77ec9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.662732] env[63293]: DEBUG oslo_concurrency.lockutils [None req-71da920f-63cc-4dcf-9e18-c2ed5df7ba22 tempest-ImagesOneServerTestJSON-2025482569 tempest-ImagesOneServerTestJSON-2025482569-project-member] Lock "c3cb139f-0e36-4044-bb8a-9deb8bc4c0f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.225s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.709726] env[63293]: DEBUG nova.network.neutron [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.990357] env[63293]: DEBUG nova.scheduler.client.report [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.006844] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquiring lock "b2c1e806-d0ae-455f-95e9-405a9557697f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.007112] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "b2c1e806-d0ae-455f-95e9-405a9557697f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.111052] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.166583] env[63293]: DEBUG nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 551.213751] env[63293]: INFO nova.compute.manager [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] [instance: aa4611e2-6254-4983-8565-899aea616726] Took 1.05 seconds to deallocate network for instance. [ 551.508183] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.708s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.508815] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 551.511189] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.686s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.595243] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "f5c47736-381f-4979-b804-494ca8cf16a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.595243] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "f5c47736-381f-4979-b804-494ca8cf16a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.697622] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.018711] env[63293]: DEBUG nova.compute.utils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 552.024120] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 552.024120] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 552.097317] env[63293]: DEBUG nova.policy [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '10bb826ae6b24969b75d457acfac9b36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8d1317599284dcf8d0c4ee422742a94', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 552.254190] env[63293]: INFO nova.scheduler.client.report [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Deleted allocations for instance aa4611e2-6254-4983-8565-899aea616726 [ 552.473073] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226f4c7f-201e-4bf2-abc0-3e9d1d028047 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.482186] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2a33f9-d80f-4f51-8b1e-e6bb1df685b3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.512517] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f400ee1-e013-4398-98c6-01ed18b2c042 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.521047] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d316c57-1648-4c2a-81eb-885fe7f1897c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.529210] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 552.542213] env[63293]: DEBUG nova.compute.provider_tree [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.567766] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Successfully created port: 51a944db-1662-4877-916e-83665a872596 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 552.761292] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e20ed6db-b84c-4f0a-820d-0034954d0c5c tempest-TenantUsagesTestJSON-436559863 tempest-TenantUsagesTestJSON-436559863-project-member] Lock "aa4611e2-6254-4983-8565-899aea616726" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.914s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.803695] env[63293]: INFO nova.compute.manager [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Rebuilding instance [ 552.876760] env[63293]: DEBUG nova.compute.manager [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 552.876760] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68243a5-224b-44c2-b3ce-d7ae882a15d9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.045202] env[63293]: DEBUG nova.scheduler.client.report [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.264169] env[63293]: DEBUG nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 553.387580] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 553.387895] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4e14b5a-97c8-4fa9-9a1d-ffce43a1eb8f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.396047] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Waiting for the task: (returnval){ [ 553.396047] env[63293]: value = "task-1327430" [ 553.396047] env[63293]: _type = "Task" [ 553.396047] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.409248] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327430, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.539220] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 553.550648] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.039s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.551386] env[63293]: ERROR nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2841f92c-3f53-4e90-a6c8-0e794b20954a, please check neutron logs for more information. [ 553.551386] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Traceback (most recent call last): [ 553.551386] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 553.551386] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] self.driver.spawn(context, instance, image_meta, [ 553.551386] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 553.551386] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.551386] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.551386] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] vm_ref = self.build_virtual_machine(instance, [ 553.551386] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.551386] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.551386] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] for vif in network_info: [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] return self._sync_wrapper(fn, *args, **kwargs) [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] self.wait() [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] self[:] = self._gt.wait() [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] return self._exit_event.wait() [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] result = hub.switch() [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.552032] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] return self.greenlet.switch() [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] result = function(*args, **kwargs) [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] return func(*args, **kwargs) [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] raise e [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] nwinfo = self.network_api.allocate_for_instance( [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] created_port_ids = self._update_ports_for_instance( [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] with excutils.save_and_reraise_exception(): [ 553.552753] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.553240] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] self.force_reraise() [ 553.553240] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.553240] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] raise self.value [ 553.553240] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.553240] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] updated_port = self._update_port( [ 553.553240] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.553240] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] _ensure_no_port_binding_failure(port) [ 553.553240] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.553240] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] raise exception.PortBindingFailed(port_id=port['id']) [ 553.553240] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] nova.exception.PortBindingFailed: Binding failed for port 2841f92c-3f53-4e90-a6c8-0e794b20954a, please check neutron logs for more information. [ 553.553240] env[63293]: ERROR nova.compute.manager [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] [ 553.553660] env[63293]: DEBUG nova.compute.utils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Binding failed for port 2841f92c-3f53-4e90-a6c8-0e794b20954a, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 553.554681] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.117s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.558024] env[63293]: DEBUG nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Build of instance c9221012-9f13-4881-9a0a-1b9c1c84d910 was re-scheduled: Binding failed for port 2841f92c-3f53-4e90-a6c8-0e794b20954a, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 553.561580] env[63293]: DEBUG nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 553.562500] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquiring lock "refresh_cache-c9221012-9f13-4881-9a0a-1b9c1c84d910" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.562500] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquired lock "refresh_cache-c9221012-9f13-4881-9a0a-1b9c1c84d910" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.562500] env[63293]: DEBUG nova.network.neutron [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 553.580587] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 553.580817] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 553.581037] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.581173] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 553.582193] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.582193] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 553.582193] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 553.582354] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 553.582395] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 553.582567] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 553.582735] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 553.583900] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b298ecc-35e0-44f7-a87c-d083c916a6da {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.594777] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1db6cd-7e31-472f-83d7-c180b6552361 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.792218] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.905548] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327430, 'name': PowerOffVM_Task, 'duration_secs': 0.122953} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.909132] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 553.909132] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 553.909132] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371ed06d-384c-4bb1-8cfa-8ac9192e50f1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.915324] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 553.915792] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5437a646-7b23-4850-83af-bba39a61efd4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.944017] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 553.944017] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 553.944017] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Deleting the datastore file [datastore1] e80209b3-273c-4045-ad8a-2a3c4c48200a {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 553.944017] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c8d2f37-4467-403e-aa87-56671ffc39b6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.950744] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Waiting for the task: (returnval){ [ 553.950744] env[63293]: value = "task-1327432" [ 553.950744] env[63293]: _type = "Task" [ 553.950744] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.958564] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327432, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.118958] env[63293]: DEBUG nova.network.neutron [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.352997] env[63293]: DEBUG nova.network.neutron [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.465128] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327432, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138919} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.469636] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 554.469636] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 554.469636] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 554.531752] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96791159-fab9-41e7-a553-e209ff42b88a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.545056] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7ad3c5-2b11-4b4d-971a-0fe49b0ecdb7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.578413] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11876f9c-2075-4c16-8e34-c90b7f4aac14 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.586093] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203eec5d-1dc3-41dc-8b2b-fd6fb0f67ac3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.600342] env[63293]: DEBUG nova.compute.provider_tree [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.860254] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Releasing lock "refresh_cache-c9221012-9f13-4881-9a0a-1b9c1c84d910" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.860254] env[63293]: DEBUG nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 554.860254] env[63293]: DEBUG nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.860571] env[63293]: DEBUG nova.network.neutron [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 554.889286] env[63293]: DEBUG nova.network.neutron [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.104907] env[63293]: DEBUG nova.scheduler.client.report [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 555.226710] env[63293]: DEBUG nova.compute.manager [req-8ced9e94-3b3b-420e-9f22-0763d64d9668 req-e5c5f39e-ca25-47cb-9662-40c3ebf71da8 service nova] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Received event network-changed-51a944db-1662-4877-916e-83665a872596 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 555.226907] env[63293]: DEBUG nova.compute.manager [req-8ced9e94-3b3b-420e-9f22-0763d64d9668 req-e5c5f39e-ca25-47cb-9662-40c3ebf71da8 service nova] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Refreshing instance network info cache due to event network-changed-51a944db-1662-4877-916e-83665a872596. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 555.227420] env[63293]: DEBUG oslo_concurrency.lockutils [req-8ced9e94-3b3b-420e-9f22-0763d64d9668 req-e5c5f39e-ca25-47cb-9662-40c3ebf71da8 service nova] Acquiring lock "refresh_cache-94ad1eb8-d583-45a1-a94e-e13eb57b2b87" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.227420] env[63293]: DEBUG oslo_concurrency.lockutils [req-8ced9e94-3b3b-420e-9f22-0763d64d9668 req-e5c5f39e-ca25-47cb-9662-40c3ebf71da8 service nova] Acquired lock "refresh_cache-94ad1eb8-d583-45a1-a94e-e13eb57b2b87" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.227420] env[63293]: DEBUG nova.network.neutron [req-8ced9e94-3b3b-420e-9f22-0763d64d9668 req-e5c5f39e-ca25-47cb-9662-40c3ebf71da8 service nova] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Refreshing network info cache for port 51a944db-1662-4877-916e-83665a872596 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 555.393769] env[63293]: DEBUG nova.network.neutron [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.409213] env[63293]: ERROR nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 51a944db-1662-4877-916e-83665a872596, please check neutron logs for more information. [ 555.409213] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 555.409213] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.409213] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 555.409213] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.409213] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 555.409213] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.409213] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 555.409213] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.409213] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 555.409213] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.409213] env[63293]: ERROR nova.compute.manager raise self.value [ 555.409213] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.409213] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 555.409213] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.409213] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 555.409777] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.409777] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 555.409777] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 51a944db-1662-4877-916e-83665a872596, please check neutron logs for more information. [ 555.409777] env[63293]: ERROR nova.compute.manager [ 555.409777] env[63293]: Traceback (most recent call last): [ 555.409777] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 555.409777] env[63293]: listener.cb(fileno) [ 555.409777] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.409777] env[63293]: result = function(*args, **kwargs) [ 555.409777] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 555.409777] env[63293]: return func(*args, **kwargs) [ 555.409777] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.409777] env[63293]: raise e [ 555.409777] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.409777] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 555.409777] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.409777] env[63293]: created_port_ids = self._update_ports_for_instance( [ 555.409777] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.409777] env[63293]: with excutils.save_and_reraise_exception(): [ 555.409777] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.409777] env[63293]: self.force_reraise() [ 555.409777] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.409777] env[63293]: raise self.value [ 555.409777] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.409777] env[63293]: updated_port = self._update_port( [ 555.409777] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.409777] env[63293]: _ensure_no_port_binding_failure(port) [ 555.409777] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.409777] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 555.410658] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 51a944db-1662-4877-916e-83665a872596, please check neutron logs for more information. [ 555.410658] env[63293]: Removing descriptor: 18 [ 555.410795] env[63293]: ERROR nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 51a944db-1662-4877-916e-83665a872596, please check neutron logs for more information. [ 555.410795] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Traceback (most recent call last): [ 555.410795] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 555.410795] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] yield resources [ 555.410795] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 555.410795] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] self.driver.spawn(context, instance, image_meta, [ 555.410795] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 555.410795] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] self._vmops.spawn(context, instance, image_meta, injected_files, [ 555.410795] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 555.410795] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] vm_ref = self.build_virtual_machine(instance, [ 555.410795] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] vif_infos = vmwarevif.get_vif_info(self._session, [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] for vif in network_info: [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] return self._sync_wrapper(fn, *args, **kwargs) [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] self.wait() [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] self[:] = self._gt.wait() [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] return self._exit_event.wait() [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 555.411933] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] result = hub.switch() [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] return self.greenlet.switch() [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] result = function(*args, **kwargs) [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] return func(*args, **kwargs) [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] raise e [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] nwinfo = self.network_api.allocate_for_instance( [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] created_port_ids = self._update_ports_for_instance( [ 555.412492] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] with excutils.save_and_reraise_exception(): [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] self.force_reraise() [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] raise self.value [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] updated_port = self._update_port( [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] _ensure_no_port_binding_failure(port) [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] raise exception.PortBindingFailed(port_id=port['id']) [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] nova.exception.PortBindingFailed: Binding failed for port 51a944db-1662-4877-916e-83665a872596, please check neutron logs for more information. [ 555.412932] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] [ 555.413306] env[63293]: INFO nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Terminating instance [ 555.413306] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "refresh_cache-94ad1eb8-d583-45a1-a94e-e13eb57b2b87" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.511572] env[63293]: DEBUG nova.virt.hardware [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 555.511572] env[63293]: DEBUG nova.virt.hardware [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 555.511572] env[63293]: DEBUG nova.virt.hardware [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 555.511901] env[63293]: DEBUG nova.virt.hardware [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 555.511901] env[63293]: DEBUG nova.virt.hardware [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 555.515227] env[63293]: DEBUG nova.virt.hardware [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 555.515227] env[63293]: DEBUG nova.virt.hardware [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 555.515227] env[63293]: DEBUG nova.virt.hardware [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 555.515227] env[63293]: DEBUG nova.virt.hardware [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 555.515227] env[63293]: DEBUG nova.virt.hardware [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 555.515564] env[63293]: DEBUG nova.virt.hardware [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 555.515564] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a272dd-1ea4-4dfe-8bf4-0577da3da49d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.523047] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae04174-4a41-4b53-b5c5-009ece505c67 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.541275] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 555.547684] env[63293]: DEBUG oslo.service.loopingcall [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.548305] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 555.548474] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3da93040-c84a-49dc-be26-083475893b49 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.566242] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 555.566242] env[63293]: value = "task-1327433" [ 555.566242] env[63293]: _type = "Task" [ 555.566242] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.575495] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327433, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.609696] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.055s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.610588] env[63293]: ERROR nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 005e83b3-4351-48e5-aefc-c6620e43ea0b, please check neutron logs for more information. [ 555.610588] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Traceback (most recent call last): [ 555.610588] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 555.610588] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] self.driver.spawn(context, instance, image_meta, [ 555.610588] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 555.610588] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 555.610588] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 555.610588] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] vm_ref = self.build_virtual_machine(instance, [ 555.610588] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 555.610588] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] vif_infos = vmwarevif.get_vif_info(self._session, [ 555.610588] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] for vif in network_info: [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] return self._sync_wrapper(fn, *args, **kwargs) [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] self.wait() [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] self[:] = self._gt.wait() [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] return self._exit_event.wait() [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] result = hub.switch() [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 555.610917] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] return self.greenlet.switch() [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] result = function(*args, **kwargs) [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] return func(*args, **kwargs) [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] raise e [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] nwinfo = self.network_api.allocate_for_instance( [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] created_port_ids = self._update_ports_for_instance( [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] with excutils.save_and_reraise_exception(): [ 555.611259] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.611764] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] self.force_reraise() [ 555.611764] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.611764] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] raise self.value [ 555.611764] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.611764] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] updated_port = self._update_port( [ 555.611764] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.611764] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] _ensure_no_port_binding_failure(port) [ 555.611764] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.611764] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] raise exception.PortBindingFailed(port_id=port['id']) [ 555.611764] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] nova.exception.PortBindingFailed: Binding failed for port 005e83b3-4351-48e5-aefc-c6620e43ea0b, please check neutron logs for more information. [ 555.611764] env[63293]: ERROR nova.compute.manager [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] [ 555.612071] env[63293]: DEBUG nova.compute.utils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Binding failed for port 005e83b3-4351-48e5-aefc-c6620e43ea0b, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 555.615781] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.238s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.619449] env[63293]: DEBUG nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Build of instance f255ca12-7110-4f73-a6ff-d6bcc122486b was re-scheduled: Binding failed for port 005e83b3-4351-48e5-aefc-c6620e43ea0b, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 555.620441] env[63293]: DEBUG nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 555.620441] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "refresh_cache-f255ca12-7110-4f73-a6ff-d6bcc122486b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.620441] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquired lock "refresh_cache-f255ca12-7110-4f73-a6ff-d6bcc122486b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.620987] env[63293]: DEBUG nova.network.neutron [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 555.773248] env[63293]: DEBUG nova.network.neutron [req-8ced9e94-3b3b-420e-9f22-0763d64d9668 req-e5c5f39e-ca25-47cb-9662-40c3ebf71da8 service nova] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.898869] env[63293]: INFO nova.compute.manager [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: c9221012-9f13-4881-9a0a-1b9c1c84d910] Took 1.04 seconds to deallocate network for instance. [ 555.939798] env[63293]: DEBUG nova.network.neutron [req-8ced9e94-3b3b-420e-9f22-0763d64d9668 req-e5c5f39e-ca25-47cb-9662-40c3ebf71da8 service nova] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.079323] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327433, 'name': CreateVM_Task, 'duration_secs': 0.299144} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.080580] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 556.080735] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.081048] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.081152] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 556.081404] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfccd365-0ceb-4e84-bb50-db96f9cbcccd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.092462] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Waiting for the task: (returnval){ [ 556.092462] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5244dd78-73f0-961a-68e5-14209ab0fc57" [ 556.092462] env[63293]: _type = "Task" [ 556.092462] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.104368] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5244dd78-73f0-961a-68e5-14209ab0fc57, 'name': SearchDatastore_Task, 'duration_secs': 0.008849} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.104368] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.104368] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 556.104368] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.104617] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.105137] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 556.105244] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53bb139e-e170-47b2-b93a-ddd5397b57be {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.119357] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 556.119357] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 556.119357] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b608eab9-b549-416f-8baf-64bf6a9576f3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.128275] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Waiting for the task: (returnval){ [ 556.128275] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52470cba-6193-8b30-0a01-f8300c4e7b1f" [ 556.128275] env[63293]: _type = "Task" [ 556.128275] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.142205] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52470cba-6193-8b30-0a01-f8300c4e7b1f, 'name': SearchDatastore_Task, 'duration_secs': 0.008557} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.143069] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b00f9b9f-b602-482c-b731-1d785cef40a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.146367] env[63293]: DEBUG nova.network.neutron [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.149591] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Waiting for the task: (returnval){ [ 556.149591] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b5a545-8a47-1dff-8db9-27a8ac6bcf77" [ 556.149591] env[63293]: _type = "Task" [ 556.149591] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.166250] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b5a545-8a47-1dff-8db9-27a8ac6bcf77, 'name': SearchDatastore_Task, 'duration_secs': 0.008101} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.168932] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.169203] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 556.169656] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-920fc569-ca2d-4106-8790-13c9093b65b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.180058] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Waiting for the task: (returnval){ [ 556.180058] env[63293]: value = "task-1327434" [ 556.180058] env[63293]: _type = "Task" [ 556.180058] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.193095] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327434, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.309431] env[63293]: DEBUG nova.network.neutron [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.443422] env[63293]: DEBUG oslo_concurrency.lockutils [req-8ced9e94-3b3b-420e-9f22-0763d64d9668 req-e5c5f39e-ca25-47cb-9662-40c3ebf71da8 service nova] Releasing lock "refresh_cache-94ad1eb8-d583-45a1-a94e-e13eb57b2b87" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.443422] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquired lock "refresh_cache-94ad1eb8-d583-45a1-a94e-e13eb57b2b87" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.446020] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 556.629945] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9014d712-67a6-4922-b135-910b8c1ebb98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.638206] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c04254-3e72-47ca-a3f7-4218d24bf263 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.671031] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0227d9-5125-4df6-90dd-eeef99fa76a2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.679397] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e48f2c-14cf-42f6-bd56-664e7b10eae8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.695193] env[63293]: DEBUG nova.compute.provider_tree [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.701555] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327434, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474113} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.701839] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 556.701922] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 556.702637] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6289340f-dcb3-47e9-b318-fc56c739f46f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.710911] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Waiting for the task: (returnval){ [ 556.710911] env[63293]: value = "task-1327435" [ 556.710911] env[63293]: _type = "Task" [ 556.710911] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.721821] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327435, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.816368] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Releasing lock "refresh_cache-f255ca12-7110-4f73-a6ff-d6bcc122486b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.819426] env[63293]: DEBUG nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 556.819623] env[63293]: DEBUG nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.819873] env[63293]: DEBUG nova.network.neutron [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 556.853898] env[63293]: DEBUG nova.network.neutron [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.947218] env[63293]: INFO nova.scheduler.client.report [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Deleted allocations for instance c9221012-9f13-4881-9a0a-1b9c1c84d910 [ 556.994415] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.199676] env[63293]: DEBUG nova.scheduler.client.report [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.223197] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327435, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065611} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.223585] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 557.224272] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fd8e7a-fbd4-4f15-80db-8735231349fd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.244598] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 557.245212] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe3a669f-66b5-472a-a2ee-16625c9bcb26 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.260999] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.268340] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Waiting for the task: (returnval){ [ 557.268340] env[63293]: value = "task-1327436" [ 557.268340] env[63293]: _type = "Task" [ 557.268340] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.277359] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327436, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.357513] env[63293]: DEBUG nova.network.neutron [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.459945] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8494ec3b-98b5-4e62-8286-907da49c5927 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "c9221012-9f13-4881-9a0a-1b9c1c84d910" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.889s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.703375] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.088s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.703873] env[63293]: ERROR nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9c1fae21-16d8-4377-8f41-e9c9154ee27f, please check neutron logs for more information. [ 557.703873] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] Traceback (most recent call last): [ 557.703873] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.703873] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] self.driver.spawn(context, instance, image_meta, [ 557.703873] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 557.703873] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.703873] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.703873] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] vm_ref = self.build_virtual_machine(instance, [ 557.703873] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.703873] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.703873] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] for vif in network_info: [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] return self._sync_wrapper(fn, *args, **kwargs) [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] self.wait() [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] self[:] = self._gt.wait() [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] return self._exit_event.wait() [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] result = hub.switch() [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.704276] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] return self.greenlet.switch() [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] result = function(*args, **kwargs) [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] return func(*args, **kwargs) [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] raise e [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] nwinfo = self.network_api.allocate_for_instance( [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] created_port_ids = self._update_ports_for_instance( [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] with excutils.save_and_reraise_exception(): [ 557.705168] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.705527] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] self.force_reraise() [ 557.705527] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.705527] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] raise self.value [ 557.705527] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.705527] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] updated_port = self._update_port( [ 557.705527] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.705527] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] _ensure_no_port_binding_failure(port) [ 557.705527] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.705527] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] raise exception.PortBindingFailed(port_id=port['id']) [ 557.705527] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] nova.exception.PortBindingFailed: Binding failed for port 9c1fae21-16d8-4377-8f41-e9c9154ee27f, please check neutron logs for more information. [ 557.705527] env[63293]: ERROR nova.compute.manager [instance: dddc6041-4436-48da-8ed8-45be063b566b] [ 557.706133] env[63293]: DEBUG nova.compute.utils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Binding failed for port 9c1fae21-16d8-4377-8f41-e9c9154ee27f, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 557.706413] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.877s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.708632] env[63293]: DEBUG nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Build of instance dddc6041-4436-48da-8ed8-45be063b566b was re-scheduled: Binding failed for port 9c1fae21-16d8-4377-8f41-e9c9154ee27f, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 557.709098] env[63293]: DEBUG nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 557.709356] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Acquiring lock "refresh_cache-dddc6041-4436-48da-8ed8-45be063b566b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.709538] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Acquired lock "refresh_cache-dddc6041-4436-48da-8ed8-45be063b566b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.709729] env[63293]: DEBUG nova.network.neutron [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.764427] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Releasing lock "refresh_cache-94ad1eb8-d583-45a1-a94e-e13eb57b2b87" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.764427] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 557.764427] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 557.764848] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-22151baf-0669-4b74-a72c-e5dea79e782b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.779856] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70daa8c-a7a8-45bf-a1ec-ba3aee1e43d2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.796524] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327436, 'name': ReconfigVM_Task, 'duration_secs': 0.307488} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.797222] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Reconfigured VM instance instance-00000002 to attach disk [datastore2] e80209b3-273c-4045-ad8a-2a3c4c48200a/e80209b3-273c-4045-ad8a-2a3c4c48200a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 557.799068] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-75f8bb2e-e492-47cb-9b55-7d6397496169 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.812397] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 94ad1eb8-d583-45a1-a94e-e13eb57b2b87 could not be found. [ 557.812560] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 557.812765] env[63293]: INFO nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Took 0.05 seconds to destroy the instance on the hypervisor. [ 557.813053] env[63293]: DEBUG oslo.service.loopingcall [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.813330] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Waiting for the task: (returnval){ [ 557.813330] env[63293]: value = "task-1327437" [ 557.813330] env[63293]: _type = "Task" [ 557.813330] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.813578] env[63293]: DEBUG nova.compute.manager [-] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.813641] env[63293]: DEBUG nova.network.neutron [-] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 557.823700] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327437, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.843536] env[63293]: DEBUG nova.network.neutron [-] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.859539] env[63293]: INFO nova.compute.manager [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: f255ca12-7110-4f73-a6ff-d6bcc122486b] Took 1.04 seconds to deallocate network for instance. [ 557.964277] env[63293]: DEBUG nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 558.052101] env[63293]: DEBUG nova.compute.manager [req-121476f2-7652-43e3-8340-8016ff83a579 req-a7691272-8f1f-403d-b02b-c84bf9e8a2cc service nova] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Received event network-vif-deleted-51a944db-1662-4877-916e-83665a872596 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 558.245716] env[63293]: DEBUG nova.network.neutron [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.330878] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327437, 'name': Rename_Task, 'duration_secs': 0.14321} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.330878] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 558.331128] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77d6a31e-9871-4e4a-a0c5-72252bc26d43 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.340437] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Waiting for the task: (returnval){ [ 558.340437] env[63293]: value = "task-1327438" [ 558.340437] env[63293]: _type = "Task" [ 558.340437] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.344377] env[63293]: DEBUG nova.network.neutron [-] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.349440] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327438, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.493192] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.498136] env[63293]: DEBUG nova.network.neutron [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.676250] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "947f69fa-77cf-4c7d-ba62-37567318b969" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.678831] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "947f69fa-77cf-4c7d-ba62-37567318b969" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.747743] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance e80209b3-273c-4045-ad8a-2a3c4c48200a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 558.851048] env[63293]: INFO nova.compute.manager [-] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Took 1.04 seconds to deallocate network for instance. [ 558.851380] env[63293]: DEBUG oslo_vmware.api [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Task: {'id': task-1327438, 'name': PowerOnVM_Task, 'duration_secs': 0.46468} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.853032] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 558.853226] env[63293]: DEBUG nova.compute.manager [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 558.853768] env[63293]: DEBUG nova.compute.claims [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 558.853884] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.854652] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da409ddc-9816-4781-9722-086bf0392ba8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.900337] env[63293]: INFO nova.scheduler.client.report [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Deleted allocations for instance f255ca12-7110-4f73-a6ff-d6bcc122486b [ 559.003496] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Releasing lock "refresh_cache-dddc6041-4436-48da-8ed8-45be063b566b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.003496] env[63293]: DEBUG nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 559.003496] env[63293]: DEBUG nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.003496] env[63293]: DEBUG nova.network.neutron [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.044497] env[63293]: DEBUG nova.network.neutron [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.253213] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance f255ca12-7110-4f73-a6ff-d6bcc122486b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.369464] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.410191] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4db7388-cf74-48ae-825b-87a1ea51c324 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "f255ca12-7110-4f73-a6ff-d6bcc122486b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.225s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.548382] env[63293]: DEBUG nova.network.neutron [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.755633] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance dddc6041-4436-48da-8ed8-45be063b566b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.755844] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 9982e59e-65af-44c7-bdd9-056565107a76 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 559.756013] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance dccd4149-a888-4910-95c8-8d0028338dee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 559.756013] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 94ad1eb8-d583-45a1-a94e-e13eb57b2b87 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 559.913972] env[63293]: DEBUG nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.051682] env[63293]: INFO nova.compute.manager [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] [instance: dddc6041-4436-48da-8ed8-45be063b566b] Took 1.05 seconds to deallocate network for instance. [ 560.263215] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 43582f6d-75fb-4734-9fb0-b71db3ddf74c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.438329] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.770589] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 0b381f4c-372f-41c8-9661-7e6bab6dd306 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.971168] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Acquiring lock "dc9ae29a-a61b-4402-a16f-460cf74bc10c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.971387] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Lock "dc9ae29a-a61b-4402-a16f-460cf74bc10c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.107094] env[63293]: INFO nova.scheduler.client.report [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Deleted allocations for instance dddc6041-4436-48da-8ed8-45be063b566b [ 561.277293] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 4dce2ec6-6f7b-4294-a32e-4cd222ae3492 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.617110] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea29a6ea-8c1d-4c57-a3bc-9b718640336c tempest-ServersTestManualDisk-667673775 tempest-ServersTestManualDisk-667673775-project-member] Lock "dddc6041-4436-48da-8ed8-45be063b566b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.674s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.783710] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 1ac2f52a-27ef-4ce4-9540-288658f08344 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.887252] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquiring lock "e80209b3-273c-4045-ad8a-2a3c4c48200a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.887538] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "e80209b3-273c-4045-ad8a-2a3c4c48200a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.887749] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquiring lock "e80209b3-273c-4045-ad8a-2a3c4c48200a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.890011] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "e80209b3-273c-4045-ad8a-2a3c4c48200a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.890011] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "e80209b3-273c-4045-ad8a-2a3c4c48200a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.891614] env[63293]: INFO nova.compute.manager [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Terminating instance [ 561.899387] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquiring lock "refresh_cache-e80209b3-273c-4045-ad8a-2a3c4c48200a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.899387] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquired lock "refresh_cache-e80209b3-273c-4045-ad8a-2a3c4c48200a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.899387] env[63293]: DEBUG nova.network.neutron [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.130299] env[63293]: DEBUG nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.288550] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance db725426-1b67-437c-aea5-cf0da50454a2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.430600] env[63293]: DEBUG nova.network.neutron [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.658745] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.693135] env[63293]: DEBUG nova.network.neutron [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.792496] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 76ceec24-eebe-4b32-b77d-f57203881a94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 563.200573] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Releasing lock "refresh_cache-e80209b3-273c-4045-ad8a-2a3c4c48200a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.200573] env[63293]: DEBUG nova.compute.manager [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 563.200573] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 563.200573] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3dd40f-74d9-4e00-b811-dc93ee7814c4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.214704] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 563.216023] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-81f8644b-1b5a-492a-8f9d-98965a07faf7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.226025] env[63293]: DEBUG oslo_vmware.api [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 563.226025] env[63293]: value = "task-1327439" [ 563.226025] env[63293]: _type = "Task" [ 563.226025] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.232185] env[63293]: DEBUG oslo_vmware.api [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327439, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.299609] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 25f0de66-914f-46b9-a15b-d936df5b3ab4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 563.740253] env[63293]: DEBUG oslo_vmware.api [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327439, 'name': PowerOffVM_Task, 'duration_secs': 0.210458} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.740253] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 563.740253] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 563.740253] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24cf4058-f1ae-498c-9ade-bd76c76fa8cf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.769257] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 563.769257] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 563.769257] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleting the datastore file [datastore2] e80209b3-273c-4045-ad8a-2a3c4c48200a {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 563.769257] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3d436f44-cffb-4e71-b827-ec3b035ba6bb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.777419] env[63293]: DEBUG oslo_vmware.api [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for the task: (returnval){ [ 563.777419] env[63293]: value = "task-1327441" [ 563.777419] env[63293]: _type = "Task" [ 563.777419] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.786430] env[63293]: DEBUG oslo_vmware.api [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327441, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.803491] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance c1f9f075-7d1c-4c05-9dfd-43820d2d07ad has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.034600] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquiring lock "92208bca-bbac-48fa-83db-3f721e6c80c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.034957] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "92208bca-bbac-48fa-83db-3f721e6c80c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.287589] env[63293]: DEBUG oslo_vmware.api [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Task: {'id': task-1327441, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097502} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.290041] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 564.290041] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 564.290041] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.290041] env[63293]: INFO nova.compute.manager [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Took 1.09 seconds to destroy the instance on the hypervisor. [ 564.290041] env[63293]: DEBUG oslo.service.loopingcall [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.290295] env[63293]: DEBUG nova.compute.manager [-] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 564.290295] env[63293]: DEBUG nova.network.neutron [-] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.308906] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance a236be17-8965-4785-834e-9ce22eeb75eb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.362993] env[63293]: DEBUG nova.network.neutron [-] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.468155] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquiring lock "305e3a9c-27cf-47b8-8825-09e5abe220ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.468537] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "305e3a9c-27cf-47b8-8825-09e5abe220ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.812694] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance a492b0a0-0bf1-4e64-bd51-ad131cd17a3b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.866505] env[63293]: DEBUG nova.network.neutron [-] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.317623] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance b2055187-edd8-4a8c-9811-183cf7ba3b39 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 565.370533] env[63293]: INFO nova.compute.manager [-] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Took 1.08 seconds to deallocate network for instance. [ 565.607236] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Acquiring lock "a63744c8-b9bf-4a89-9319-061d53de6b5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.608992] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Lock "a63744c8-b9bf-4a89-9319-061d53de6b5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.822300] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 510b9ea6-04f1-4704-b5e4-e19db57769a8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 565.881256] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.327862] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance eaaf2515-cc4b-42c7-aacd-14074c760924 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 566.833114] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 1814ebc5-7869-4ec6-8cf6-790daed5e271 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 566.975807] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Acquiring lock "8b4b041b-5d3f-479a-8241-c75cce650d1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.976219] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Lock "8b4b041b-5d3f-479a-8241-c75cce650d1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.307144] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquiring lock "8c2439e1-26ce-45e5-9179-164c0dbbbf76" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.307603] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "8c2439e1-26ce-45e5-9179-164c0dbbbf76" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.337458] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 567.841047] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 87c7bb2c-46ff-45f7-8967-14192f2b42e4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.056921] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Acquiring lock "e0e08728-b6c3-4610-bc27-348c5594fefc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.057241] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Lock "e0e08728-b6c3-4610-bc27-348c5594fefc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.345367] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 769c2643-57eb-456c-8083-997a8c9c48cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.850407] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance b2c1e806-d0ae-455f-95e9-405a9557697f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 569.353628] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance f5c47736-381f-4979-b804-494ca8cf16a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 569.540069] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Acquiring lock "2afc50fb-138c-4523-b086-db11026ae52f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.540326] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Lock "2afc50fb-138c-4523-b086-db11026ae52f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.858853] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 947f69fa-77cf-4c7d-ba62-37567318b969 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 569.859567] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 569.859567] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 570.341430] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01445b3-a396-4662-b40e-968b15926e0f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.356357] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27da196-b677-4566-ba59-ef87bd354a46 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.392889] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d9314f-a248-400d-8626-465ceef707b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.401705] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7fc74b-7464-47a3-ac8b-dce0631a632c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.416375] env[63293]: DEBUG nova.compute.provider_tree [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.900144] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Acquiring lock "5bf150a5-ad49-4bf8-a8fa-e376e0f00926" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.900372] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Lock "5bf150a5-ad49-4bf8-a8fa-e376e0f00926" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.920302] env[63293]: DEBUG nova.scheduler.client.report [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.428160] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 571.428440] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.722s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.428819] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.562s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.408742] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd27a67f-76f6-4546-9707-a78879c1755e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.416926] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-340f4ab6-98c4-418b-99f0-f9054f241f95 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.448933] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ef56c9-abb3-4d0d-81ba-1e9641ebe019 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.458941] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8423d66b-7ba9-40f9-b3ee-d16f08c01269 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.472866] env[63293]: DEBUG nova.compute.provider_tree [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.976860] env[63293]: DEBUG nova.scheduler.client.report [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.483987] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.055s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.484677] env[63293]: ERROR nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f5b367dd-acd0-41f7-81bf-cfdcdafe0e82, please check neutron logs for more information. [ 573.484677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Traceback (most recent call last): [ 573.484677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.484677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] self.driver.spawn(context, instance, image_meta, [ 573.484677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 573.484677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.484677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.484677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] vm_ref = self.build_virtual_machine(instance, [ 573.484677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.484677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.484677] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] for vif in network_info: [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] return self._sync_wrapper(fn, *args, **kwargs) [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] self.wait() [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] self[:] = self._gt.wait() [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] return self._exit_event.wait() [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] result = hub.switch() [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.485114] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] return self.greenlet.switch() [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] result = function(*args, **kwargs) [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] return func(*args, **kwargs) [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] raise e [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] nwinfo = self.network_api.allocate_for_instance( [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] created_port_ids = self._update_ports_for_instance( [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] with excutils.save_and_reraise_exception(): [ 573.485523] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.485938] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] self.force_reraise() [ 573.485938] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.485938] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] raise self.value [ 573.485938] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.485938] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] updated_port = self._update_port( [ 573.485938] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.485938] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] _ensure_no_port_binding_failure(port) [ 573.485938] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.485938] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] raise exception.PortBindingFailed(port_id=port['id']) [ 573.485938] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] nova.exception.PortBindingFailed: Binding failed for port f5b367dd-acd0-41f7-81bf-cfdcdafe0e82, please check neutron logs for more information. [ 573.485938] env[63293]: ERROR nova.compute.manager [instance: 9982e59e-65af-44c7-bdd9-056565107a76] [ 573.486251] env[63293]: DEBUG nova.compute.utils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Binding failed for port f5b367dd-acd0-41f7-81bf-cfdcdafe0e82, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 573.487075] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.497s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.491356] env[63293]: DEBUG nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Build of instance 9982e59e-65af-44c7-bdd9-056565107a76 was re-scheduled: Binding failed for port f5b367dd-acd0-41f7-81bf-cfdcdafe0e82, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 573.492260] env[63293]: DEBUG nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 573.492677] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquiring lock "refresh_cache-9982e59e-65af-44c7-bdd9-056565107a76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.492785] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquired lock "refresh_cache-9982e59e-65af-44c7-bdd9-056565107a76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.492970] env[63293]: DEBUG nova.network.neutron [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.032640] env[63293]: DEBUG nova.network.neutron [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.159141] env[63293]: DEBUG nova.network.neutron [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.606419] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d461f07-b8e8-4b5b-86b3-580593874ef1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.616415] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20003342-df16-4840-92f5-d4da9dd0cb2a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.648780] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca67de2-0a59-44b2-b1c5-6095787f9711 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.656722] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9be61c4-43a1-4697-9f42-cb52b5115a17 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.672283] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Releasing lock "refresh_cache-9982e59e-65af-44c7-bdd9-056565107a76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.672572] env[63293]: DEBUG nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 574.672703] env[63293]: DEBUG nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.672870] env[63293]: DEBUG nova.network.neutron [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.675610] env[63293]: DEBUG nova.compute.provider_tree [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.704152] env[63293]: DEBUG nova.network.neutron [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.178312] env[63293]: DEBUG nova.scheduler.client.report [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.208649] env[63293]: DEBUG nova.network.neutron [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.684237] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.197s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.685025] env[63293]: ERROR nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2b781a20-f80f-43a9-aebd-2fe700fc0ae4, please check neutron logs for more information. [ 575.685025] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] Traceback (most recent call last): [ 575.685025] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.685025] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] self.driver.spawn(context, instance, image_meta, [ 575.685025] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 575.685025] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.685025] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.685025] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] vm_ref = self.build_virtual_machine(instance, [ 575.685025] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.685025] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.685025] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] for vif in network_info: [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] return self._sync_wrapper(fn, *args, **kwargs) [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] self.wait() [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] self[:] = self._gt.wait() [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] return self._exit_event.wait() [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] result = hub.switch() [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.686558] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] return self.greenlet.switch() [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] result = function(*args, **kwargs) [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] return func(*args, **kwargs) [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] raise e [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] nwinfo = self.network_api.allocate_for_instance( [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] created_port_ids = self._update_ports_for_instance( [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] with excutils.save_and_reraise_exception(): [ 575.686970] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.687336] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] self.force_reraise() [ 575.687336] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.687336] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] raise self.value [ 575.687336] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.687336] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] updated_port = self._update_port( [ 575.687336] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.687336] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] _ensure_no_port_binding_failure(port) [ 575.687336] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.687336] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] raise exception.PortBindingFailed(port_id=port['id']) [ 575.687336] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] nova.exception.PortBindingFailed: Binding failed for port 2b781a20-f80f-43a9-aebd-2fe700fc0ae4, please check neutron logs for more information. [ 575.687336] env[63293]: ERROR nova.compute.manager [instance: dccd4149-a888-4910-95c8-8d0028338dee] [ 575.687645] env[63293]: DEBUG nova.compute.utils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Binding failed for port 2b781a20-f80f-43a9-aebd-2fe700fc0ae4, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 575.689884] env[63293]: DEBUG nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Build of instance dccd4149-a888-4910-95c8-8d0028338dee was re-scheduled: Binding failed for port 2b781a20-f80f-43a9-aebd-2fe700fc0ae4, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 575.690372] env[63293]: DEBUG nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 575.690610] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "refresh_cache-dccd4149-a888-4910-95c8-8d0028338dee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.690757] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquired lock "refresh_cache-dccd4149-a888-4910-95c8-8d0028338dee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.690919] env[63293]: DEBUG nova.network.neutron [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 575.697391] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.050s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.697391] env[63293]: INFO nova.compute.claims [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.711073] env[63293]: INFO nova.compute.manager [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 9982e59e-65af-44c7-bdd9-056565107a76] Took 1.04 seconds to deallocate network for instance. [ 575.859120] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "e86185ea-7aa6-43ad-920d-645e18194b04" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.859378] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "e86185ea-7aa6-43ad-920d-645e18194b04" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.213361] env[63293]: DEBUG nova.network.neutron [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.376033] env[63293]: DEBUG nova.network.neutron [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.752285] env[63293]: INFO nova.scheduler.client.report [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Deleted allocations for instance 9982e59e-65af-44c7-bdd9-056565107a76 [ 576.793466] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquiring lock "0aade4da-882a-4342-aefb-07eb306b17ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.793702] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lock "0aade4da-882a-4342-aefb-07eb306b17ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.878167] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Releasing lock "refresh_cache-dccd4149-a888-4910-95c8-8d0028338dee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.878395] env[63293]: DEBUG nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 576.878575] env[63293]: DEBUG nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.878740] env[63293]: DEBUG nova.network.neutron [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 576.898130] env[63293]: DEBUG nova.network.neutron [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.189580] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62eaf4c8-e8f3-4e9c-8a6b-0bbf192e141e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.197679] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a470d04-d3c1-4ef7-8b87-9d3bcc6ee961 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.226637] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7add14-cadd-4137-88e1-8cee308db3f9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.234801] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce4f3135-d6c9-46e8-9e49-e9bc3eaec9b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.251534] env[63293]: DEBUG nova.compute.provider_tree [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.264921] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6c86674a-d60a-4a08-8d26-94721eb9ebd8 tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "9982e59e-65af-44c7-bdd9-056565107a76" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.357s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.402293] env[63293]: DEBUG nova.network.neutron [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.757188] env[63293]: DEBUG nova.scheduler.client.report [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.768346] env[63293]: DEBUG nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.905548] env[63293]: INFO nova.compute.manager [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: dccd4149-a888-4910-95c8-8d0028338dee] Took 1.03 seconds to deallocate network for instance. [ 578.264541] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.570s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.264541] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 578.266681] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.726s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.268023] env[63293]: INFO nova.compute.claims [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 578.295322] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.774106] env[63293]: DEBUG nova.compute.utils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 578.784503] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 578.784503] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 578.846710] env[63293]: DEBUG nova.policy [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '10bb826ae6b24969b75d457acfac9b36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8d1317599284dcf8d0c4ee422742a94', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 578.948217] env[63293]: INFO nova.scheduler.client.report [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Deleted allocations for instance dccd4149-a888-4910-95c8-8d0028338dee [ 579.240268] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Successfully created port: cb3b734d-a172-403d-a199-800fe83a10ea {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.286993] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 579.458876] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd7c1a88-9119-47f9-8f85-8a6a04ce2274 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "dccd4149-a888-4910-95c8-8d0028338dee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.722s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.850282] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d562af6-c321-4f73-9574-83b73bc4b498 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.858528] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7020a7-bd03-4455-998c-fe8ac9b096d6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.893764] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179e07e5-8955-4a10-b2d7-ced0dfca19c3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.900991] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3762067b-9d04-470c-96dc-e5be68230088 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.914411] env[63293]: DEBUG nova.compute.provider_tree [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.960864] env[63293]: DEBUG nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.300338] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 580.339687] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 580.339935] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 580.340102] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 580.340279] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 580.340420] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 580.340566] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 580.340768] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 580.340927] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 580.341100] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 580.341261] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 580.341426] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 580.342454] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07be3039-3c48-484d-8450-813c171f427d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.354343] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9caa4d55-d884-46c2-babe-a0506831e370 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.418259] env[63293]: DEBUG nova.scheduler.client.report [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.485745] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.538054] env[63293]: DEBUG nova.compute.manager [req-308286fd-bcde-463c-963a-f5019f0ab96f req-20404a1f-ee23-4e88-8a3b-76a0fd150c36 service nova] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Received event network-changed-cb3b734d-a172-403d-a199-800fe83a10ea {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.538252] env[63293]: DEBUG nova.compute.manager [req-308286fd-bcde-463c-963a-f5019f0ab96f req-20404a1f-ee23-4e88-8a3b-76a0fd150c36 service nova] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Refreshing instance network info cache due to event network-changed-cb3b734d-a172-403d-a199-800fe83a10ea. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 580.538457] env[63293]: DEBUG oslo_concurrency.lockutils [req-308286fd-bcde-463c-963a-f5019f0ab96f req-20404a1f-ee23-4e88-8a3b-76a0fd150c36 service nova] Acquiring lock "refresh_cache-43582f6d-75fb-4734-9fb0-b71db3ddf74c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.538594] env[63293]: DEBUG oslo_concurrency.lockutils [req-308286fd-bcde-463c-963a-f5019f0ab96f req-20404a1f-ee23-4e88-8a3b-76a0fd150c36 service nova] Acquired lock "refresh_cache-43582f6d-75fb-4734-9fb0-b71db3ddf74c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.538749] env[63293]: DEBUG nova.network.neutron [req-308286fd-bcde-463c-963a-f5019f0ab96f req-20404a1f-ee23-4e88-8a3b-76a0fd150c36 service nova] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Refreshing network info cache for port cb3b734d-a172-403d-a199-800fe83a10ea {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 580.581338] env[63293]: ERROR nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cb3b734d-a172-403d-a199-800fe83a10ea, please check neutron logs for more information. [ 580.581338] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 580.581338] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.581338] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 580.581338] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 580.581338] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 580.581338] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 580.581338] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 580.581338] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.581338] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 580.581338] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.581338] env[63293]: ERROR nova.compute.manager raise self.value [ 580.581338] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 580.581338] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 580.581338] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.581338] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 580.581896] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.581896] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 580.581896] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cb3b734d-a172-403d-a199-800fe83a10ea, please check neutron logs for more information. [ 580.581896] env[63293]: ERROR nova.compute.manager [ 580.581896] env[63293]: Traceback (most recent call last): [ 580.581896] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 580.581896] env[63293]: listener.cb(fileno) [ 580.581896] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.581896] env[63293]: result = function(*args, **kwargs) [ 580.581896] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 580.581896] env[63293]: return func(*args, **kwargs) [ 580.581896] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.581896] env[63293]: raise e [ 580.581896] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.581896] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 580.581896] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 580.581896] env[63293]: created_port_ids = self._update_ports_for_instance( [ 580.581896] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 580.581896] env[63293]: with excutils.save_and_reraise_exception(): [ 580.581896] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.581896] env[63293]: self.force_reraise() [ 580.581896] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.581896] env[63293]: raise self.value [ 580.581896] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 580.581896] env[63293]: updated_port = self._update_port( [ 580.581896] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.581896] env[63293]: _ensure_no_port_binding_failure(port) [ 580.581896] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.581896] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 580.582831] env[63293]: nova.exception.PortBindingFailed: Binding failed for port cb3b734d-a172-403d-a199-800fe83a10ea, please check neutron logs for more information. [ 580.582831] env[63293]: Removing descriptor: 18 [ 580.582831] env[63293]: ERROR nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cb3b734d-a172-403d-a199-800fe83a10ea, please check neutron logs for more information. [ 580.582831] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Traceback (most recent call last): [ 580.582831] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 580.582831] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] yield resources [ 580.582831] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 580.582831] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] self.driver.spawn(context, instance, image_meta, [ 580.582831] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 580.582831] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.582831] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.582831] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] vm_ref = self.build_virtual_machine(instance, [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] for vif in network_info: [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] return self._sync_wrapper(fn, *args, **kwargs) [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] self.wait() [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] self[:] = self._gt.wait() [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] return self._exit_event.wait() [ 580.583218] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] result = hub.switch() [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] return self.greenlet.switch() [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] result = function(*args, **kwargs) [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] return func(*args, **kwargs) [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] raise e [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] nwinfo = self.network_api.allocate_for_instance( [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 580.583608] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] created_port_ids = self._update_ports_for_instance( [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] with excutils.save_and_reraise_exception(): [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] self.force_reraise() [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] raise self.value [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] updated_port = self._update_port( [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] _ensure_no_port_binding_failure(port) [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.584058] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] raise exception.PortBindingFailed(port_id=port['id']) [ 580.584426] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] nova.exception.PortBindingFailed: Binding failed for port cb3b734d-a172-403d-a199-800fe83a10ea, please check neutron logs for more information. [ 580.584426] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] [ 580.584426] env[63293]: INFO nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Terminating instance [ 580.585536] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "refresh_cache-43582f6d-75fb-4734-9fb0-b71db3ddf74c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.923863] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.657s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.924425] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 580.927018] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.817s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.927290] env[63293]: DEBUG nova.objects.instance [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63293) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 581.058054] env[63293]: DEBUG nova.network.neutron [req-308286fd-bcde-463c-963a-f5019f0ab96f req-20404a1f-ee23-4e88-8a3b-76a0fd150c36 service nova] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.153965] env[63293]: DEBUG nova.network.neutron [req-308286fd-bcde-463c-963a-f5019f0ab96f req-20404a1f-ee23-4e88-8a3b-76a0fd150c36 service nova] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.432391] env[63293]: DEBUG nova.compute.utils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 581.438111] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 581.438111] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 581.489616] env[63293]: DEBUG nova.policy [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '10bb826ae6b24969b75d457acfac9b36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8d1317599284dcf8d0c4ee422742a94', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 581.658459] env[63293]: DEBUG oslo_concurrency.lockutils [req-308286fd-bcde-463c-963a-f5019f0ab96f req-20404a1f-ee23-4e88-8a3b-76a0fd150c36 service nova] Releasing lock "refresh_cache-43582f6d-75fb-4734-9fb0-b71db3ddf74c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.659112] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquired lock "refresh_cache-43582f6d-75fb-4734-9fb0-b71db3ddf74c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.659349] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.872991] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Successfully created port: 60ec81ef-8266-4005-aef5-f24d13bcf605 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 581.937831] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 581.941767] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b1608e7-a383-4025-b7b4-b12b58ba739f tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.942696] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.245s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.946444] env[63293]: INFO nova.compute.claims [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.184755] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.322753] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.565825] env[63293]: DEBUG nova.compute.manager [req-b69745dd-5312-401c-8a2e-55336e78e2de req-9a5c46d5-bad1-44be-9c25-50269dc9b4c9 service nova] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Received event network-vif-deleted-cb3b734d-a172-403d-a199-800fe83a10ea {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 582.825567] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Releasing lock "refresh_cache-43582f6d-75fb-4734-9fb0-b71db3ddf74c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.825990] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 582.826197] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 582.826948] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67aab6b2-8420-45a1-9d7a-44ddf74f2fe5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.837437] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11fa6fc5-4a37-453c-8461-34909269b0fc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.858790] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 43582f6d-75fb-4734-9fb0-b71db3ddf74c could not be found. [ 582.859072] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 582.859431] env[63293]: INFO nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 582.859431] env[63293]: DEBUG oslo.service.loopingcall [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.859648] env[63293]: DEBUG nova.compute.manager [-] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.859743] env[63293]: DEBUG nova.network.neutron [-] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 582.888834] env[63293]: DEBUG nova.network.neutron [-] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.953148] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 582.986982] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 582.987473] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 582.987799] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.992017] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 582.992017] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.992017] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 582.992017] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 582.992017] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 582.993106] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 582.993106] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 582.993106] env[63293]: DEBUG nova.virt.hardware [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 582.993844] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5885f2a-e89e-4efb-a3f1-570c4c2bbabd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.002385] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d70038e-8c91-482b-8a23-0f6e0ceff728 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.148274] env[63293]: ERROR nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 60ec81ef-8266-4005-aef5-f24d13bcf605, please check neutron logs for more information. [ 583.148274] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 583.148274] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.148274] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 583.148274] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.148274] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 583.148274] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.148274] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 583.148274] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.148274] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 583.148274] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.148274] env[63293]: ERROR nova.compute.manager raise self.value [ 583.148274] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.148274] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 583.148274] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.148274] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 583.149010] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.149010] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 583.149010] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 60ec81ef-8266-4005-aef5-f24d13bcf605, please check neutron logs for more information. [ 583.149010] env[63293]: ERROR nova.compute.manager [ 583.149010] env[63293]: Traceback (most recent call last): [ 583.149010] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 583.149010] env[63293]: listener.cb(fileno) [ 583.149010] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.149010] env[63293]: result = function(*args, **kwargs) [ 583.149010] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 583.149010] env[63293]: return func(*args, **kwargs) [ 583.149010] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.149010] env[63293]: raise e [ 583.149010] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.149010] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 583.149010] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.149010] env[63293]: created_port_ids = self._update_ports_for_instance( [ 583.149010] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.149010] env[63293]: with excutils.save_and_reraise_exception(): [ 583.149010] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.149010] env[63293]: self.force_reraise() [ 583.149010] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.149010] env[63293]: raise self.value [ 583.149010] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.149010] env[63293]: updated_port = self._update_port( [ 583.149010] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.149010] env[63293]: _ensure_no_port_binding_failure(port) [ 583.149010] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.149010] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 583.149992] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 60ec81ef-8266-4005-aef5-f24d13bcf605, please check neutron logs for more information. [ 583.149992] env[63293]: Removing descriptor: 18 [ 583.149992] env[63293]: ERROR nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 60ec81ef-8266-4005-aef5-f24d13bcf605, please check neutron logs for more information. [ 583.149992] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Traceback (most recent call last): [ 583.149992] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 583.149992] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] yield resources [ 583.149992] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 583.149992] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] self.driver.spawn(context, instance, image_meta, [ 583.149992] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 583.149992] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.149992] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.149992] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] vm_ref = self.build_virtual_machine(instance, [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] for vif in network_info: [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] return self._sync_wrapper(fn, *args, **kwargs) [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] self.wait() [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] self[:] = self._gt.wait() [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] return self._exit_event.wait() [ 583.150369] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] result = hub.switch() [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] return self.greenlet.switch() [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] result = function(*args, **kwargs) [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] return func(*args, **kwargs) [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] raise e [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] nwinfo = self.network_api.allocate_for_instance( [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.150769] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] created_port_ids = self._update_ports_for_instance( [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] with excutils.save_and_reraise_exception(): [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] self.force_reraise() [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] raise self.value [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] updated_port = self._update_port( [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] _ensure_no_port_binding_failure(port) [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.151259] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] raise exception.PortBindingFailed(port_id=port['id']) [ 583.151611] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] nova.exception.PortBindingFailed: Binding failed for port 60ec81ef-8266-4005-aef5-f24d13bcf605, please check neutron logs for more information. [ 583.151611] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] [ 583.151611] env[63293]: INFO nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Terminating instance [ 583.155188] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "refresh_cache-0b381f4c-372f-41c8-9661-7e6bab6dd306" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.155188] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquired lock "refresh_cache-0b381f4c-372f-41c8-9661-7e6bab6dd306" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.155188] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.391458] env[63293]: DEBUG nova.network.neutron [-] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.430940] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a3824ac-d3d6-4535-8f00-173fc10addf7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.438760] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c24a1a-9fd9-4b21-94f0-a2e7591fb3d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.471834] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff930821-c29b-486a-9a1a-91435eb7cc92 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.480323] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7186b1c-a2c2-43ae-8a58-c97643646cf3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.497306] env[63293]: DEBUG nova.compute.provider_tree [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.679311] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.783255] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.900677] env[63293]: INFO nova.compute.manager [-] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Took 1.04 seconds to deallocate network for instance. [ 583.903349] env[63293]: DEBUG nova.compute.claims [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 583.903603] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.000935] env[63293]: DEBUG nova.scheduler.client.report [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.286715] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Releasing lock "refresh_cache-0b381f4c-372f-41c8-9661-7e6bab6dd306" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.287210] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 584.287472] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 584.287819] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24038e17-6313-4435-a886-82c7b15be4ff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.297071] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f66ce95-370c-43af-9055-07827deefce5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.322068] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0b381f4c-372f-41c8-9661-7e6bab6dd306 could not be found. [ 584.322306] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 584.322489] env[63293]: INFO nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Took 0.04 seconds to destroy the instance on the hypervisor. [ 584.322729] env[63293]: DEBUG oslo.service.loopingcall [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.322946] env[63293]: DEBUG nova.compute.manager [-] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 584.323070] env[63293]: DEBUG nova.network.neutron [-] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.342767] env[63293]: DEBUG nova.network.neutron [-] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.511519] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.512376] env[63293]: DEBUG nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 584.515623] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.724s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.519096] env[63293]: INFO nova.compute.claims [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 584.589151] env[63293]: DEBUG nova.compute.manager [req-495e9c5b-f002-4246-8bf0-a62d1295c807 req-ce0e8351-da31-4212-8f39-235126238f5e service nova] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Received event network-changed-60ec81ef-8266-4005-aef5-f24d13bcf605 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 584.589151] env[63293]: DEBUG nova.compute.manager [req-495e9c5b-f002-4246-8bf0-a62d1295c807 req-ce0e8351-da31-4212-8f39-235126238f5e service nova] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Refreshing instance network info cache due to event network-changed-60ec81ef-8266-4005-aef5-f24d13bcf605. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 584.589250] env[63293]: DEBUG oslo_concurrency.lockutils [req-495e9c5b-f002-4246-8bf0-a62d1295c807 req-ce0e8351-da31-4212-8f39-235126238f5e service nova] Acquiring lock "refresh_cache-0b381f4c-372f-41c8-9661-7e6bab6dd306" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.589818] env[63293]: DEBUG oslo_concurrency.lockutils [req-495e9c5b-f002-4246-8bf0-a62d1295c807 req-ce0e8351-da31-4212-8f39-235126238f5e service nova] Acquired lock "refresh_cache-0b381f4c-372f-41c8-9661-7e6bab6dd306" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.589818] env[63293]: DEBUG nova.network.neutron [req-495e9c5b-f002-4246-8bf0-a62d1295c807 req-ce0e8351-da31-4212-8f39-235126238f5e service nova] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Refreshing network info cache for port 60ec81ef-8266-4005-aef5-f24d13bcf605 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 584.848687] env[63293]: DEBUG nova.network.neutron [-] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.022617] env[63293]: DEBUG nova.compute.utils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.024037] env[63293]: DEBUG nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 585.030016] env[63293]: DEBUG nova.network.neutron [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 585.084734] env[63293]: DEBUG nova.policy [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd2a9495ab0b4768a4075a71f25aa9a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff8479fc251544e9b4618deccf992754', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.113889] env[63293]: DEBUG nova.network.neutron [req-495e9c5b-f002-4246-8bf0-a62d1295c807 req-ce0e8351-da31-4212-8f39-235126238f5e service nova] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.263028] env[63293]: DEBUG nova.network.neutron [req-495e9c5b-f002-4246-8bf0-a62d1295c807 req-ce0e8351-da31-4212-8f39-235126238f5e service nova] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.351572] env[63293]: INFO nova.compute.manager [-] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Took 1.03 seconds to deallocate network for instance. [ 585.354048] env[63293]: DEBUG nova.compute.claims [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 585.354217] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.490894] env[63293]: DEBUG nova.network.neutron [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Successfully created port: 0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 585.538773] env[63293]: DEBUG nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 585.764846] env[63293]: DEBUG oslo_concurrency.lockutils [req-495e9c5b-f002-4246-8bf0-a62d1295c807 req-ce0e8351-da31-4212-8f39-235126238f5e service nova] Releasing lock "refresh_cache-0b381f4c-372f-41c8-9661-7e6bab6dd306" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.765131] env[63293]: DEBUG nova.compute.manager [req-495e9c5b-f002-4246-8bf0-a62d1295c807 req-ce0e8351-da31-4212-8f39-235126238f5e service nova] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Received event network-vif-deleted-60ec81ef-8266-4005-aef5-f24d13bcf605 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.054750] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a783e466-dd44-4e53-94f0-2645b5a94102 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.067697] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6f5a5c-216b-4266-a4a6-b64db7a98dc5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.101317] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20503e7e-ffc0-4e61-89ac-df5f2dd3c659 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.110154] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6b530c-84d6-47a9-a4f9-019a32a26e4a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.124406] env[63293]: DEBUG nova.compute.provider_tree [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.560228] env[63293]: DEBUG nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 586.594332] env[63293]: DEBUG nova.virt.hardware [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 586.594332] env[63293]: DEBUG nova.virt.hardware [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 586.594332] env[63293]: DEBUG nova.virt.hardware [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 586.594474] env[63293]: DEBUG nova.virt.hardware [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 586.594474] env[63293]: DEBUG nova.virt.hardware [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 586.594474] env[63293]: DEBUG nova.virt.hardware [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 586.594474] env[63293]: DEBUG nova.virt.hardware [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 586.594474] env[63293]: DEBUG nova.virt.hardware [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 586.594658] env[63293]: DEBUG nova.virt.hardware [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 586.594658] env[63293]: DEBUG nova.virt.hardware [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 586.594658] env[63293]: DEBUG nova.virt.hardware [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 586.594658] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56028dee-1617-49fd-95e1-35923adc2113 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.604012] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d2eae0-2dff-46ad-a1c9-9669de7d11aa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.627973] env[63293]: DEBUG nova.scheduler.client.report [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.724154] env[63293]: DEBUG nova.compute.manager [req-85d46dcd-f1bf-43bf-a609-91689be69fd7 req-b23ec14c-b802-407d-a7fa-39c4efd7a3c0 service nova] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Received event network-changed-0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.724414] env[63293]: DEBUG nova.compute.manager [req-85d46dcd-f1bf-43bf-a609-91689be69fd7 req-b23ec14c-b802-407d-a7fa-39c4efd7a3c0 service nova] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Refreshing instance network info cache due to event network-changed-0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 586.724639] env[63293]: DEBUG oslo_concurrency.lockutils [req-85d46dcd-f1bf-43bf-a609-91689be69fd7 req-b23ec14c-b802-407d-a7fa-39c4efd7a3c0 service nova] Acquiring lock "refresh_cache-4dce2ec6-6f7b-4294-a32e-4cd222ae3492" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.724874] env[63293]: DEBUG oslo_concurrency.lockutils [req-85d46dcd-f1bf-43bf-a609-91689be69fd7 req-b23ec14c-b802-407d-a7fa-39c4efd7a3c0 service nova] Acquired lock "refresh_cache-4dce2ec6-6f7b-4294-a32e-4cd222ae3492" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.725597] env[63293]: DEBUG nova.network.neutron [req-85d46dcd-f1bf-43bf-a609-91689be69fd7 req-b23ec14c-b802-407d-a7fa-39c4efd7a3c0 service nova] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Refreshing network info cache for port 0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 586.733772] env[63293]: ERROR nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c, please check neutron logs for more information. [ 586.733772] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 586.733772] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.733772] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 586.733772] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.733772] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 586.733772] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.733772] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 586.733772] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.733772] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 586.733772] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.733772] env[63293]: ERROR nova.compute.manager raise self.value [ 586.733772] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.733772] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 586.733772] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.733772] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 586.734285] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.734285] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 586.734285] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c, please check neutron logs for more information. [ 586.734285] env[63293]: ERROR nova.compute.manager [ 586.734285] env[63293]: Traceback (most recent call last): [ 586.734285] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 586.734285] env[63293]: listener.cb(fileno) [ 586.734285] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.734285] env[63293]: result = function(*args, **kwargs) [ 586.734285] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 586.734285] env[63293]: return func(*args, **kwargs) [ 586.734285] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.734285] env[63293]: raise e [ 586.734285] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.734285] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 586.734285] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.734285] env[63293]: created_port_ids = self._update_ports_for_instance( [ 586.734285] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.734285] env[63293]: with excutils.save_and_reraise_exception(): [ 586.734285] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.734285] env[63293]: self.force_reraise() [ 586.734285] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.734285] env[63293]: raise self.value [ 586.734285] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.734285] env[63293]: updated_port = self._update_port( [ 586.734285] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.734285] env[63293]: _ensure_no_port_binding_failure(port) [ 586.734285] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.734285] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 586.735152] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c, please check neutron logs for more information. [ 586.735152] env[63293]: Removing descriptor: 18 [ 586.735152] env[63293]: ERROR nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c, please check neutron logs for more information. [ 586.735152] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Traceback (most recent call last): [ 586.735152] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 586.735152] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] yield resources [ 586.735152] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.735152] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] self.driver.spawn(context, instance, image_meta, [ 586.735152] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 586.735152] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.735152] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.735152] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] vm_ref = self.build_virtual_machine(instance, [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] for vif in network_info: [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] return self._sync_wrapper(fn, *args, **kwargs) [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] self.wait() [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] self[:] = self._gt.wait() [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] return self._exit_event.wait() [ 586.735548] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] result = hub.switch() [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] return self.greenlet.switch() [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] result = function(*args, **kwargs) [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] return func(*args, **kwargs) [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] raise e [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] nwinfo = self.network_api.allocate_for_instance( [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.735971] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] created_port_ids = self._update_ports_for_instance( [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] with excutils.save_and_reraise_exception(): [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] self.force_reraise() [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] raise self.value [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] updated_port = self._update_port( [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] _ensure_no_port_binding_failure(port) [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.736398] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] raise exception.PortBindingFailed(port_id=port['id']) [ 586.736851] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] nova.exception.PortBindingFailed: Binding failed for port 0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c, please check neutron logs for more information. [ 586.736851] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] [ 586.736851] env[63293]: INFO nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Terminating instance [ 586.738827] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-4dce2ec6-6f7b-4294-a32e-4cd222ae3492" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.905573] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "db04bc40-1106-46dd-ad9a-f9b11a6c8774" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.905778] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "db04bc40-1106-46dd-ad9a-f9b11a6c8774" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.133273] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.618s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.133807] env[63293]: DEBUG nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 587.136412] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.643s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.137899] env[63293]: INFO nova.compute.claims [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.245071] env[63293]: DEBUG nova.network.neutron [req-85d46dcd-f1bf-43bf-a609-91689be69fd7 req-b23ec14c-b802-407d-a7fa-39c4efd7a3c0 service nova] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.328277] env[63293]: DEBUG nova.network.neutron [req-85d46dcd-f1bf-43bf-a609-91689be69fd7 req-b23ec14c-b802-407d-a7fa-39c4efd7a3c0 service nova] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.643981] env[63293]: DEBUG nova.compute.utils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 587.648201] env[63293]: DEBUG nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 587.648201] env[63293]: DEBUG nova.network.neutron [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 587.702740] env[63293]: DEBUG nova.policy [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b905d104776482cb262ee41c0638433', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '29986647349c4ea6a7f0bc878d5161b0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 587.830991] env[63293]: DEBUG oslo_concurrency.lockutils [req-85d46dcd-f1bf-43bf-a609-91689be69fd7 req-b23ec14c-b802-407d-a7fa-39c4efd7a3c0 service nova] Releasing lock "refresh_cache-4dce2ec6-6f7b-4294-a32e-4cd222ae3492" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.831446] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-4dce2ec6-6f7b-4294-a32e-4cd222ae3492" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.831630] env[63293]: DEBUG nova.network.neutron [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.004600] env[63293]: DEBUG nova.network.neutron [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Successfully created port: 9567745f-7393-4661-994a-fb9813f6b138 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.147368] env[63293]: DEBUG nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 588.359829] env[63293]: DEBUG nova.network.neutron [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.520292] env[63293]: DEBUG nova.network.neutron [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.635942] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa7f0d2-2cb5-44da-8b57-e588a7fa39c9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.647498] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262eb228-870c-4ba8-bb3d-615250a475a2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.689247] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08792039-2fa7-47d8-874e-5671a2448000 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.696683] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf52283-2adb-4466-bf31-5db5e28870bb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.710436] env[63293]: DEBUG nova.compute.provider_tree [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.791385] env[63293]: DEBUG nova.compute.manager [req-3c24ac53-7e24-46fb-82f5-55116a1af949 req-bf418adf-d51c-4ecc-9ba8-f7fe34577cf1 service nova] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Received event network-vif-deleted-0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.024388] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-4dce2ec6-6f7b-4294-a32e-4cd222ae3492" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.024857] env[63293]: DEBUG nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.025074] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.025382] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b564d1b8-fdd8-480a-a4ad-887529ea6fc0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.034417] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166068cd-0a89-4600-8d8c-763366618057 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.055366] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4dce2ec6-6f7b-4294-a32e-4cd222ae3492 could not be found. [ 589.055563] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 589.055740] env[63293]: INFO nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Took 0.03 seconds to destroy the instance on the hypervisor. [ 589.055975] env[63293]: DEBUG oslo.service.loopingcall [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.056193] env[63293]: DEBUG nova.compute.manager [-] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.056288] env[63293]: DEBUG nova.network.neutron [-] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.075490] env[63293]: DEBUG nova.network.neutron [-] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.111111] env[63293]: ERROR nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9567745f-7393-4661-994a-fb9813f6b138, please check neutron logs for more information. [ 589.111111] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.111111] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.111111] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.111111] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.111111] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.111111] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.111111] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.111111] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.111111] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 589.111111] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.111111] env[63293]: ERROR nova.compute.manager raise self.value [ 589.111111] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.111111] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.111111] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.111111] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.111671] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.111671] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.111671] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9567745f-7393-4661-994a-fb9813f6b138, please check neutron logs for more information. [ 589.111671] env[63293]: ERROR nova.compute.manager [ 589.111671] env[63293]: Traceback (most recent call last): [ 589.111671] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.111671] env[63293]: listener.cb(fileno) [ 589.111671] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.111671] env[63293]: result = function(*args, **kwargs) [ 589.111671] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.111671] env[63293]: return func(*args, **kwargs) [ 589.111671] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.111671] env[63293]: raise e [ 589.111671] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.111671] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 589.111671] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.111671] env[63293]: created_port_ids = self._update_ports_for_instance( [ 589.111671] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.111671] env[63293]: with excutils.save_and_reraise_exception(): [ 589.111671] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.111671] env[63293]: self.force_reraise() [ 589.111671] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.111671] env[63293]: raise self.value [ 589.111671] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.111671] env[63293]: updated_port = self._update_port( [ 589.111671] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.111671] env[63293]: _ensure_no_port_binding_failure(port) [ 589.111671] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.111671] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.112552] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 9567745f-7393-4661-994a-fb9813f6b138, please check neutron logs for more information. [ 589.112552] env[63293]: Removing descriptor: 18 [ 589.190603] env[63293]: DEBUG nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 589.214640] env[63293]: DEBUG nova.virt.hardware [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 589.214878] env[63293]: DEBUG nova.virt.hardware [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 589.215048] env[63293]: DEBUG nova.virt.hardware [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.215230] env[63293]: DEBUG nova.virt.hardware [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 589.215372] env[63293]: DEBUG nova.virt.hardware [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.215516] env[63293]: DEBUG nova.virt.hardware [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 589.215715] env[63293]: DEBUG nova.virt.hardware [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 589.215882] env[63293]: DEBUG nova.virt.hardware [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 589.216060] env[63293]: DEBUG nova.virt.hardware [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 589.216226] env[63293]: DEBUG nova.virt.hardware [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 589.216396] env[63293]: DEBUG nova.virt.hardware [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.218106] env[63293]: DEBUG nova.scheduler.client.report [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.221082] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5ffa16-9e1d-4cad-baa5-c2c3c2ef2dae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.230036] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f46cf0-16e8-4521-b1a0-0387fc3abeee {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.244059] env[63293]: ERROR nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9567745f-7393-4661-994a-fb9813f6b138, please check neutron logs for more information. [ 589.244059] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Traceback (most recent call last): [ 589.244059] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 589.244059] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] yield resources [ 589.244059] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.244059] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] self.driver.spawn(context, instance, image_meta, [ 589.244059] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 589.244059] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.244059] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.244059] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] vm_ref = self.build_virtual_machine(instance, [ 589.244059] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] for vif in network_info: [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] return self._sync_wrapper(fn, *args, **kwargs) [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] self.wait() [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] self[:] = self._gt.wait() [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] return self._exit_event.wait() [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 589.244553] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] current.throw(*self._exc) [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] result = function(*args, **kwargs) [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] return func(*args, **kwargs) [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] raise e [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] nwinfo = self.network_api.allocate_for_instance( [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] created_port_ids = self._update_ports_for_instance( [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] with excutils.save_and_reraise_exception(): [ 589.244976] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.245485] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] self.force_reraise() [ 589.245485] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.245485] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] raise self.value [ 589.245485] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.245485] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] updated_port = self._update_port( [ 589.245485] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.245485] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] _ensure_no_port_binding_failure(port) [ 589.245485] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.245485] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] raise exception.PortBindingFailed(port_id=port['id']) [ 589.245485] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] nova.exception.PortBindingFailed: Binding failed for port 9567745f-7393-4661-994a-fb9813f6b138, please check neutron logs for more information. [ 589.245485] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] [ 589.245485] env[63293]: INFO nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Terminating instance [ 589.247502] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquiring lock "refresh_cache-1ac2f52a-27ef-4ce4-9540-288658f08344" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.247654] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquired lock "refresh_cache-1ac2f52a-27ef-4ce4-9540-288658f08344" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.247835] env[63293]: DEBUG nova.network.neutron [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.577574] env[63293]: DEBUG nova.network.neutron [-] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.725785] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.726249] env[63293]: DEBUG nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 589.730279] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.876s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.765183] env[63293]: DEBUG nova.network.neutron [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.862126] env[63293]: DEBUG nova.network.neutron [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.080565] env[63293]: INFO nova.compute.manager [-] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Took 1.02 seconds to deallocate network for instance. [ 590.082826] env[63293]: DEBUG nova.compute.claims [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.083006] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.236674] env[63293]: DEBUG nova.compute.utils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 590.241399] env[63293]: DEBUG nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 590.241399] env[63293]: DEBUG nova.network.neutron [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 590.278847] env[63293]: DEBUG nova.policy [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3bb00827293c4b74a6f86f8db3a09cfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9169d8b45ed148b8935d1f8ca96ac3e0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 590.365755] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Releasing lock "refresh_cache-1ac2f52a-27ef-4ce4-9540-288658f08344" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.366646] env[63293]: DEBUG nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 590.366646] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 590.366899] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c42f77d8-63f6-4893-ace9-46e481069a75 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.383357] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31573dc5-27e1-4269-95e9-9ca1d65d2ffe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.412064] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1ac2f52a-27ef-4ce4-9540-288658f08344 could not be found. [ 590.412064] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 590.412064] env[63293]: INFO nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Took 0.05 seconds to destroy the instance on the hypervisor. [ 590.412322] env[63293]: DEBUG oslo.service.loopingcall [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.414589] env[63293]: DEBUG nova.compute.manager [-] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.414705] env[63293]: DEBUG nova.network.neutron [-] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.488687] env[63293]: DEBUG nova.network.neutron [-] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.668151] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-362f66cb-4cde-42b3-8455-06d1bb0a2c78 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.675497] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4070b3-0086-4b13-853d-99fa6969a445 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.708164] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8afe04-7484-4822-bb4a-bb77ba5a247e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.716027] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefaf7f3-2e01-47d6-a70f-c6f39185274a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.729723] env[63293]: DEBUG nova.compute.provider_tree [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.741531] env[63293]: DEBUG nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 590.753792] env[63293]: DEBUG nova.network.neutron [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Successfully created port: d6936f7c-3207-4cca-92d7-bdc13912b1ab {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.836642] env[63293]: DEBUG nova.compute.manager [req-f989031b-04f5-4356-9666-d22a2405c1b3 req-ff10c1e3-4502-4237-9a01-19cb9faf6379 service nova] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Received event network-changed-9567745f-7393-4661-994a-fb9813f6b138 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 590.836695] env[63293]: DEBUG nova.compute.manager [req-f989031b-04f5-4356-9666-d22a2405c1b3 req-ff10c1e3-4502-4237-9a01-19cb9faf6379 service nova] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Refreshing instance network info cache due to event network-changed-9567745f-7393-4661-994a-fb9813f6b138. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 590.836927] env[63293]: DEBUG oslo_concurrency.lockutils [req-f989031b-04f5-4356-9666-d22a2405c1b3 req-ff10c1e3-4502-4237-9a01-19cb9faf6379 service nova] Acquiring lock "refresh_cache-1ac2f52a-27ef-4ce4-9540-288658f08344" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.837095] env[63293]: DEBUG oslo_concurrency.lockutils [req-f989031b-04f5-4356-9666-d22a2405c1b3 req-ff10c1e3-4502-4237-9a01-19cb9faf6379 service nova] Acquired lock "refresh_cache-1ac2f52a-27ef-4ce4-9540-288658f08344" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.837264] env[63293]: DEBUG nova.network.neutron [req-f989031b-04f5-4356-9666-d22a2405c1b3 req-ff10c1e3-4502-4237-9a01-19cb9faf6379 service nova] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Refreshing network info cache for port 9567745f-7393-4661-994a-fb9813f6b138 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 590.991696] env[63293]: DEBUG nova.network.neutron [-] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.234113] env[63293]: DEBUG nova.scheduler.client.report [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.358189] env[63293]: DEBUG nova.network.neutron [req-f989031b-04f5-4356-9666-d22a2405c1b3 req-ff10c1e3-4502-4237-9a01-19cb9faf6379 service nova] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.467398] env[63293]: DEBUG nova.network.neutron [req-f989031b-04f5-4356-9666-d22a2405c1b3 req-ff10c1e3-4502-4237-9a01-19cb9faf6379 service nova] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.495986] env[63293]: INFO nova.compute.manager [-] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Took 1.08 seconds to deallocate network for instance. [ 591.498219] env[63293]: DEBUG nova.compute.claims [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 591.498418] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.742017] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.009s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.742017] env[63293]: ERROR nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 51a944db-1662-4877-916e-83665a872596, please check neutron logs for more information. [ 591.742017] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Traceback (most recent call last): [ 591.742017] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.742017] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] self.driver.spawn(context, instance, image_meta, [ 591.742017] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 591.742017] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.742017] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.742017] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] vm_ref = self.build_virtual_machine(instance, [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] for vif in network_info: [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] return self._sync_wrapper(fn, *args, **kwargs) [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] self.wait() [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] self[:] = self._gt.wait() [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] return self._exit_event.wait() [ 591.742390] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] result = hub.switch() [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] return self.greenlet.switch() [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] result = function(*args, **kwargs) [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] return func(*args, **kwargs) [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] raise e [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] nwinfo = self.network_api.allocate_for_instance( [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.742763] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] created_port_ids = self._update_ports_for_instance( [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] with excutils.save_and_reraise_exception(): [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] self.force_reraise() [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] raise self.value [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] updated_port = self._update_port( [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] _ensure_no_port_binding_failure(port) [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.743146] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] raise exception.PortBindingFailed(port_id=port['id']) [ 591.743487] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] nova.exception.PortBindingFailed: Binding failed for port 51a944db-1662-4877-916e-83665a872596, please check neutron logs for more information. [ 591.743487] env[63293]: ERROR nova.compute.manager [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] [ 591.743487] env[63293]: DEBUG nova.compute.utils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Binding failed for port 51a944db-1662-4877-916e-83665a872596, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 591.744123] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Build of instance 94ad1eb8-d583-45a1-a94e-e13eb57b2b87 was re-scheduled: Binding failed for port 51a944db-1662-4877-916e-83665a872596, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 591.745329] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 591.745329] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "refresh_cache-94ad1eb8-d583-45a1-a94e-e13eb57b2b87" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.745329] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquired lock "refresh_cache-94ad1eb8-d583-45a1-a94e-e13eb57b2b87" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.745329] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.746839] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 32.377s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.746839] env[63293]: DEBUG nova.objects.instance [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63293) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 591.753601] env[63293]: DEBUG nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 591.797774] env[63293]: DEBUG nova.virt.hardware [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:32:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.798066] env[63293]: DEBUG nova.virt.hardware [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.798263] env[63293]: DEBUG nova.virt.hardware [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.798406] env[63293]: DEBUG nova.virt.hardware [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.798548] env[63293]: DEBUG nova.virt.hardware [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.799371] env[63293]: DEBUG nova.virt.hardware [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.799371] env[63293]: DEBUG nova.virt.hardware [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.803016] env[63293]: DEBUG nova.virt.hardware [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.803016] env[63293]: DEBUG nova.virt.hardware [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.803016] env[63293]: DEBUG nova.virt.hardware [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.803016] env[63293]: DEBUG nova.virt.hardware [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.803016] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392303a1-9507-4eb3-817a-857f84c3ab12 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.812431] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6d2193-2a4d-4bcb-8b0b-493b21335635 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.970031] env[63293]: DEBUG oslo_concurrency.lockutils [req-f989031b-04f5-4356-9666-d22a2405c1b3 req-ff10c1e3-4502-4237-9a01-19cb9faf6379 service nova] Releasing lock "refresh_cache-1ac2f52a-27ef-4ce4-9540-288658f08344" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.970323] env[63293]: DEBUG nova.compute.manager [req-f989031b-04f5-4356-9666-d22a2405c1b3 req-ff10c1e3-4502-4237-9a01-19cb9faf6379 service nova] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Received event network-vif-deleted-9567745f-7393-4661-994a-fb9813f6b138 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.015193] env[63293]: ERROR nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d6936f7c-3207-4cca-92d7-bdc13912b1ab, please check neutron logs for more information. [ 592.015193] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 592.015193] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.015193] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 592.015193] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.015193] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 592.015193] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.015193] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 592.015193] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.015193] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 592.015193] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.015193] env[63293]: ERROR nova.compute.manager raise self.value [ 592.015193] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.015193] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 592.015193] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.015193] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 592.015697] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.015697] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 592.015697] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d6936f7c-3207-4cca-92d7-bdc13912b1ab, please check neutron logs for more information. [ 592.015697] env[63293]: ERROR nova.compute.manager [ 592.015697] env[63293]: Traceback (most recent call last): [ 592.015697] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 592.015697] env[63293]: listener.cb(fileno) [ 592.015697] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.015697] env[63293]: result = function(*args, **kwargs) [ 592.015697] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.015697] env[63293]: return func(*args, **kwargs) [ 592.015697] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.015697] env[63293]: raise e [ 592.015697] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.015697] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 592.015697] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.015697] env[63293]: created_port_ids = self._update_ports_for_instance( [ 592.015697] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.015697] env[63293]: with excutils.save_and_reraise_exception(): [ 592.015697] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.015697] env[63293]: self.force_reraise() [ 592.015697] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.015697] env[63293]: raise self.value [ 592.015697] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.015697] env[63293]: updated_port = self._update_port( [ 592.015697] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.015697] env[63293]: _ensure_no_port_binding_failure(port) [ 592.015697] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.015697] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 592.016537] env[63293]: nova.exception.PortBindingFailed: Binding failed for port d6936f7c-3207-4cca-92d7-bdc13912b1ab, please check neutron logs for more information. [ 592.016537] env[63293]: Removing descriptor: 18 [ 592.016537] env[63293]: ERROR nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d6936f7c-3207-4cca-92d7-bdc13912b1ab, please check neutron logs for more information. [ 592.016537] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] Traceback (most recent call last): [ 592.016537] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 592.016537] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] yield resources [ 592.016537] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.016537] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] self.driver.spawn(context, instance, image_meta, [ 592.016537] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 592.016537] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.016537] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.016537] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] vm_ref = self.build_virtual_machine(instance, [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] for vif in network_info: [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] return self._sync_wrapper(fn, *args, **kwargs) [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] self.wait() [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] self[:] = self._gt.wait() [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] return self._exit_event.wait() [ 592.016961] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] result = hub.switch() [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] return self.greenlet.switch() [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] result = function(*args, **kwargs) [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] return func(*args, **kwargs) [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] raise e [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] nwinfo = self.network_api.allocate_for_instance( [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.017367] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] created_port_ids = self._update_ports_for_instance( [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] with excutils.save_and_reraise_exception(): [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] self.force_reraise() [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] raise self.value [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] updated_port = self._update_port( [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] _ensure_no_port_binding_failure(port) [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.017876] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] raise exception.PortBindingFailed(port_id=port['id']) [ 592.018631] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] nova.exception.PortBindingFailed: Binding failed for port d6936f7c-3207-4cca-92d7-bdc13912b1ab, please check neutron logs for more information. [ 592.018631] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] [ 592.018631] env[63293]: INFO nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Terminating instance [ 592.019257] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "refresh_cache-db725426-1b67-437c-aea5-cf0da50454a2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.020090] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquired lock "refresh_cache-db725426-1b67-437c-aea5-cf0da50454a2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.020308] env[63293]: DEBUG nova.network.neutron [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 592.269467] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.348575] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.537691] env[63293]: DEBUG nova.network.neutron [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.614690] env[63293]: DEBUG nova.network.neutron [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.758379] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f65bd848-0dd8-4274-8e9d-ac76faec2ef9 tempest-ServersAdmin275Test-1900682338 tempest-ServersAdmin275Test-1900682338-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.759579] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.322s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.761023] env[63293]: INFO nova.compute.claims [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.853060] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Releasing lock "refresh_cache-94ad1eb8-d583-45a1-a94e-e13eb57b2b87" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.853060] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 592.853060] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.853060] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.858638] env[63293]: DEBUG nova.compute.manager [req-d5f4e414-4cda-4512-afe5-8cc13caa7217 req-e9ab73db-7048-4350-8f7e-51b850a0d0a7 service nova] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Received event network-changed-d6936f7c-3207-4cca-92d7-bdc13912b1ab {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.858823] env[63293]: DEBUG nova.compute.manager [req-d5f4e414-4cda-4512-afe5-8cc13caa7217 req-e9ab73db-7048-4350-8f7e-51b850a0d0a7 service nova] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Refreshing instance network info cache due to event network-changed-d6936f7c-3207-4cca-92d7-bdc13912b1ab. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 592.859027] env[63293]: DEBUG oslo_concurrency.lockutils [req-d5f4e414-4cda-4512-afe5-8cc13caa7217 req-e9ab73db-7048-4350-8f7e-51b850a0d0a7 service nova] Acquiring lock "refresh_cache-db725426-1b67-437c-aea5-cf0da50454a2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.878375] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.117157] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Releasing lock "refresh_cache-db725426-1b67-437c-aea5-cf0da50454a2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.117583] env[63293]: DEBUG nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 593.117777] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 593.118090] env[63293]: DEBUG oslo_concurrency.lockutils [req-d5f4e414-4cda-4512-afe5-8cc13caa7217 req-e9ab73db-7048-4350-8f7e-51b850a0d0a7 service nova] Acquired lock "refresh_cache-db725426-1b67-437c-aea5-cf0da50454a2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.118260] env[63293]: DEBUG nova.network.neutron [req-d5f4e414-4cda-4512-afe5-8cc13caa7217 req-e9ab73db-7048-4350-8f7e-51b850a0d0a7 service nova] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Refreshing network info cache for port d6936f7c-3207-4cca-92d7-bdc13912b1ab {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 593.119979] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-794c7fa8-f790-4dcb-a161-85e57758bc14 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.130703] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d771a291-662e-434c-bab2-195a4c1921bc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.151567] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance db725426-1b67-437c-aea5-cf0da50454a2 could not be found. [ 593.151769] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 593.151944] env[63293]: INFO nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 593.152202] env[63293]: DEBUG oslo.service.loopingcall [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.152415] env[63293]: DEBUG nova.compute.manager [-] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.152505] env[63293]: DEBUG nova.network.neutron [-] [instance: db725426-1b67-437c-aea5-cf0da50454a2] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.166320] env[63293]: DEBUG nova.network.neutron [-] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.385135] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.645386] env[63293]: DEBUG nova.network.neutron [req-d5f4e414-4cda-4512-afe5-8cc13caa7217 req-e9ab73db-7048-4350-8f7e-51b850a0d0a7 service nova] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.668817] env[63293]: DEBUG nova.network.neutron [-] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.734609] env[63293]: DEBUG nova.network.neutron [req-d5f4e414-4cda-4512-afe5-8cc13caa7217 req-e9ab73db-7048-4350-8f7e-51b850a0d0a7 service nova] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.888696] env[63293]: INFO nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 94ad1eb8-d583-45a1-a94e-e13eb57b2b87] Took 1.04 seconds to deallocate network for instance. [ 594.172769] env[63293]: INFO nova.compute.manager [-] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Took 1.02 seconds to deallocate network for instance. [ 594.174954] env[63293]: DEBUG nova.compute.claims [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 594.175140] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.237285] env[63293]: DEBUG oslo_concurrency.lockutils [req-d5f4e414-4cda-4512-afe5-8cc13caa7217 req-e9ab73db-7048-4350-8f7e-51b850a0d0a7 service nova] Releasing lock "refresh_cache-db725426-1b67-437c-aea5-cf0da50454a2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.237546] env[63293]: DEBUG nova.compute.manager [req-d5f4e414-4cda-4512-afe5-8cc13caa7217 req-e9ab73db-7048-4350-8f7e-51b850a0d0a7 service nova] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Received event network-vif-deleted-d6936f7c-3207-4cca-92d7-bdc13912b1ab {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.239271] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73cf4233-7762-4b7b-bd3c-9bdd65e343f2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.246673] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af10f995-6f6f-45f5-bff2-f1b66b2e5cca {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.275736] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc23f42-bdf4-4004-a8a6-7c0565e6ab2c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.282686] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283a95c8-f9b4-4ca3-86be-ba0005d68dd9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.295329] env[63293]: DEBUG nova.compute.provider_tree [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.798423] env[63293]: DEBUG nova.scheduler.client.report [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.916685] env[63293]: INFO nova.scheduler.client.report [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Deleted allocations for instance 94ad1eb8-d583-45a1-a94e-e13eb57b2b87 [ 595.303513] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.303935] env[63293]: DEBUG nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.306585] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.648s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.308067] env[63293]: INFO nova.compute.claims [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.430820] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "94ad1eb8-d583-45a1-a94e-e13eb57b2b87" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.644s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.814166] env[63293]: DEBUG nova.compute.utils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.818572] env[63293]: DEBUG nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.818808] env[63293]: DEBUG nova.network.neutron [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 595.869449] env[63293]: DEBUG nova.policy [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e97b276ffcac4cffaa88852005858cf6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '755d0ab8a85247db97ace5251db90567', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.933761] env[63293]: DEBUG nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 596.174661] env[63293]: DEBUG nova.network.neutron [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Successfully created port: 53b6765d-312f-412f-9291-ac729b337036 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.319285] env[63293]: DEBUG nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.459551] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.779976] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f7e083e-02ce-4887-adaa-1a8e050ef062 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.787510] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaffa6f3-b5f9-42bd-bdd8-6628592f2616 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.816835] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6cf110-7eba-4ff5-98ba-ed6b80b4411d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.824196] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdea5b3a-0cf9-474a-85d2-66eb3d167706 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.839904] env[63293]: DEBUG nova.compute.provider_tree [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.051360] env[63293]: DEBUG nova.compute.manager [req-afc0b6e5-c239-453a-9e33-5c57ecb95dd8 req-960241a5-6c1a-456b-9bd3-5e47da6d0859 service nova] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Received event network-changed-53b6765d-312f-412f-9291-ac729b337036 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 597.051558] env[63293]: DEBUG nova.compute.manager [req-afc0b6e5-c239-453a-9e33-5c57ecb95dd8 req-960241a5-6c1a-456b-9bd3-5e47da6d0859 service nova] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Refreshing instance network info cache due to event network-changed-53b6765d-312f-412f-9291-ac729b337036. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 597.051777] env[63293]: DEBUG oslo_concurrency.lockutils [req-afc0b6e5-c239-453a-9e33-5c57ecb95dd8 req-960241a5-6c1a-456b-9bd3-5e47da6d0859 service nova] Acquiring lock "refresh_cache-76ceec24-eebe-4b32-b77d-f57203881a94" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.051919] env[63293]: DEBUG oslo_concurrency.lockutils [req-afc0b6e5-c239-453a-9e33-5c57ecb95dd8 req-960241a5-6c1a-456b-9bd3-5e47da6d0859 service nova] Acquired lock "refresh_cache-76ceec24-eebe-4b32-b77d-f57203881a94" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.057018] env[63293]: DEBUG nova.network.neutron [req-afc0b6e5-c239-453a-9e33-5c57ecb95dd8 req-960241a5-6c1a-456b-9bd3-5e47da6d0859 service nova] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Refreshing network info cache for port 53b6765d-312f-412f-9291-ac729b337036 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 597.243962] env[63293]: ERROR nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 53b6765d-312f-412f-9291-ac729b337036, please check neutron logs for more information. [ 597.243962] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.243962] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.243962] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.243962] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.243962] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.243962] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.243962] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.243962] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.243962] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 597.243962] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.243962] env[63293]: ERROR nova.compute.manager raise self.value [ 597.243962] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.243962] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.243962] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.243962] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.244539] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.244539] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.244539] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 53b6765d-312f-412f-9291-ac729b337036, please check neutron logs for more information. [ 597.244539] env[63293]: ERROR nova.compute.manager [ 597.244539] env[63293]: Traceback (most recent call last): [ 597.244539] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.244539] env[63293]: listener.cb(fileno) [ 597.244539] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.244539] env[63293]: result = function(*args, **kwargs) [ 597.244539] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 597.244539] env[63293]: return func(*args, **kwargs) [ 597.244539] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.244539] env[63293]: raise e [ 597.244539] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.244539] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 597.244539] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.244539] env[63293]: created_port_ids = self._update_ports_for_instance( [ 597.244539] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.244539] env[63293]: with excutils.save_and_reraise_exception(): [ 597.244539] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.244539] env[63293]: self.force_reraise() [ 597.244539] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.244539] env[63293]: raise self.value [ 597.244539] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.244539] env[63293]: updated_port = self._update_port( [ 597.244539] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.244539] env[63293]: _ensure_no_port_binding_failure(port) [ 597.244539] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.244539] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.245524] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 53b6765d-312f-412f-9291-ac729b337036, please check neutron logs for more information. [ 597.245524] env[63293]: Removing descriptor: 18 [ 597.342677] env[63293]: DEBUG nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.347438] env[63293]: DEBUG nova.scheduler.client.report [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.373570] env[63293]: DEBUG nova.virt.hardware [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.373810] env[63293]: DEBUG nova.virt.hardware [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.373966] env[63293]: DEBUG nova.virt.hardware [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.374166] env[63293]: DEBUG nova.virt.hardware [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.374311] env[63293]: DEBUG nova.virt.hardware [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.374457] env[63293]: DEBUG nova.virt.hardware [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.374663] env[63293]: DEBUG nova.virt.hardware [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.374824] env[63293]: DEBUG nova.virt.hardware [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.375030] env[63293]: DEBUG nova.virt.hardware [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.375202] env[63293]: DEBUG nova.virt.hardware [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.375395] env[63293]: DEBUG nova.virt.hardware [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.376516] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d66273-9f32-44e5-b462-b84d948b3982 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.385929] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a2336d-eb4a-4425-8035-a58af901908c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.399030] env[63293]: ERROR nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 53b6765d-312f-412f-9291-ac729b337036, please check neutron logs for more information. [ 597.399030] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Traceback (most recent call last): [ 597.399030] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 597.399030] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] yield resources [ 597.399030] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.399030] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] self.driver.spawn(context, instance, image_meta, [ 597.399030] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 597.399030] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.399030] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.399030] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] vm_ref = self.build_virtual_machine(instance, [ 597.399030] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] for vif in network_info: [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] return self._sync_wrapper(fn, *args, **kwargs) [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] self.wait() [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] self[:] = self._gt.wait() [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] return self._exit_event.wait() [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 597.399490] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] current.throw(*self._exc) [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] result = function(*args, **kwargs) [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] return func(*args, **kwargs) [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] raise e [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] nwinfo = self.network_api.allocate_for_instance( [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] created_port_ids = self._update_ports_for_instance( [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] with excutils.save_and_reraise_exception(): [ 597.399927] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.400370] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] self.force_reraise() [ 597.400370] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.400370] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] raise self.value [ 597.400370] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.400370] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] updated_port = self._update_port( [ 597.400370] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.400370] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] _ensure_no_port_binding_failure(port) [ 597.400370] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.400370] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] raise exception.PortBindingFailed(port_id=port['id']) [ 597.400370] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] nova.exception.PortBindingFailed: Binding failed for port 53b6765d-312f-412f-9291-ac729b337036, please check neutron logs for more information. [ 597.400370] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] [ 597.400370] env[63293]: INFO nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Terminating instance [ 597.426622] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Acquiring lock "refresh_cache-76ceec24-eebe-4b32-b77d-f57203881a94" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.569201] env[63293]: DEBUG nova.network.neutron [req-afc0b6e5-c239-453a-9e33-5c57ecb95dd8 req-960241a5-6c1a-456b-9bd3-5e47da6d0859 service nova] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.664841] env[63293]: DEBUG nova.network.neutron [req-afc0b6e5-c239-453a-9e33-5c57ecb95dd8 req-960241a5-6c1a-456b-9bd3-5e47da6d0859 service nova] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.853445] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.853958] env[63293]: DEBUG nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 597.856447] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.976s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.856669] env[63293]: DEBUG nova.objects.instance [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lazy-loading 'resources' on Instance uuid e80209b3-273c-4045-ad8a-2a3c4c48200a {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 598.167652] env[63293]: DEBUG oslo_concurrency.lockutils [req-afc0b6e5-c239-453a-9e33-5c57ecb95dd8 req-960241a5-6c1a-456b-9bd3-5e47da6d0859 service nova] Releasing lock "refresh_cache-76ceec24-eebe-4b32-b77d-f57203881a94" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.168082] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Acquired lock "refresh_cache-76ceec24-eebe-4b32-b77d-f57203881a94" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.168248] env[63293]: DEBUG nova.network.neutron [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.359581] env[63293]: DEBUG nova.compute.utils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.360934] env[63293]: DEBUG nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 598.361126] env[63293]: DEBUG nova.network.neutron [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 598.402652] env[63293]: DEBUG nova.policy [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b0819e2b6ce4bd0a7ec7edf5cc409b3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a0b610289cb8435797f8e9b3cafe962f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 598.706140] env[63293]: DEBUG nova.network.neutron [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.752287] env[63293]: DEBUG nova.network.neutron [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Successfully created port: c54cc217-14c7-43a7-9f9c-a980c39413ee {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 598.807873] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4bc251b-d75b-4624-8923-1b905f1d1db9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.815720] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78182ac-32d3-4757-a750-b4b115bfbe07 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.853324] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d35629-371f-4d1a-9af6-7bd7a0f7caea {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.861379] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19d7342-38a3-4d6b-a677-a0b927e77652 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.866355] env[63293]: DEBUG nova.network.neutron [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.867704] env[63293]: DEBUG nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 598.880244] env[63293]: DEBUG nova.compute.provider_tree [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.182938] env[63293]: DEBUG nova.compute.manager [req-02861f18-ffcb-4673-ad53-898dd5522e47 req-ed7818ca-00d9-4f1a-8095-7f8c1ce16281 service nova] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Received event network-vif-deleted-53b6765d-312f-412f-9291-ac729b337036 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 599.373156] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Releasing lock "refresh_cache-76ceec24-eebe-4b32-b77d-f57203881a94" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.373156] env[63293]: DEBUG nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.373156] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 599.376128] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-78a93850-9a61-40b3-bbd5-ace8f177ac01 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.382862] env[63293]: DEBUG nova.scheduler.client.report [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.390221] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da82ec5-e656-43a4-9564-981fd9c9c454 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.414331] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 76ceec24-eebe-4b32-b77d-f57203881a94 could not be found. [ 599.415334] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 599.415334] env[63293]: INFO nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Took 0.04 seconds to destroy the instance on the hypervisor. [ 599.415334] env[63293]: DEBUG oslo.service.loopingcall [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.415513] env[63293]: DEBUG nova.compute.manager [-] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.415685] env[63293]: DEBUG nova.network.neutron [-] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.438446] env[63293]: DEBUG nova.network.neutron [-] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.880142] env[63293]: DEBUG nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 599.891113] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.031s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.891113] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.595s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.892139] env[63293]: INFO nova.compute.claims [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.908144] env[63293]: DEBUG nova.virt.hardware [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 599.908144] env[63293]: DEBUG nova.virt.hardware [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 599.908338] env[63293]: DEBUG nova.virt.hardware [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.908394] env[63293]: DEBUG nova.virt.hardware [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 599.908497] env[63293]: DEBUG nova.virt.hardware [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.908670] env[63293]: DEBUG nova.virt.hardware [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 599.908841] env[63293]: DEBUG nova.virt.hardware [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 599.908998] env[63293]: DEBUG nova.virt.hardware [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 599.909176] env[63293]: DEBUG nova.virt.hardware [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 599.909345] env[63293]: DEBUG nova.virt.hardware [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 599.909513] env[63293]: DEBUG nova.virt.hardware [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.910968] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db41b5a-d69f-4424-85f0-ab481648c2a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.915910] env[63293]: INFO nova.scheduler.client.report [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Deleted allocations for instance e80209b3-273c-4045-ad8a-2a3c4c48200a [ 599.923664] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65fdde42-7810-450f-9bfd-9cacb0adb716 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.940559] env[63293]: DEBUG nova.network.neutron [-] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.978850] env[63293]: ERROR nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c54cc217-14c7-43a7-9f9c-a980c39413ee, please check neutron logs for more information. [ 599.978850] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 599.978850] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.978850] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 599.978850] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.978850] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 599.978850] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.978850] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 599.978850] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.978850] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 599.978850] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.978850] env[63293]: ERROR nova.compute.manager raise self.value [ 599.978850] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.978850] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 599.978850] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.978850] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 599.979583] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.979583] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 599.979583] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c54cc217-14c7-43a7-9f9c-a980c39413ee, please check neutron logs for more information. [ 599.979583] env[63293]: ERROR nova.compute.manager [ 599.979583] env[63293]: Traceback (most recent call last): [ 599.979583] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 599.979583] env[63293]: listener.cb(fileno) [ 599.979583] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.979583] env[63293]: result = function(*args, **kwargs) [ 599.979583] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 599.979583] env[63293]: return func(*args, **kwargs) [ 599.979583] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.979583] env[63293]: raise e [ 599.979583] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.979583] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 599.979583] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.979583] env[63293]: created_port_ids = self._update_ports_for_instance( [ 599.979583] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.979583] env[63293]: with excutils.save_and_reraise_exception(): [ 599.979583] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.979583] env[63293]: self.force_reraise() [ 599.979583] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.979583] env[63293]: raise self.value [ 599.979583] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.979583] env[63293]: updated_port = self._update_port( [ 599.979583] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.979583] env[63293]: _ensure_no_port_binding_failure(port) [ 599.979583] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.979583] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 599.980602] env[63293]: nova.exception.PortBindingFailed: Binding failed for port c54cc217-14c7-43a7-9f9c-a980c39413ee, please check neutron logs for more information. [ 599.980602] env[63293]: Removing descriptor: 18 [ 599.980602] env[63293]: ERROR nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c54cc217-14c7-43a7-9f9c-a980c39413ee, please check neutron logs for more information. [ 599.980602] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Traceback (most recent call last): [ 599.980602] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 599.980602] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] yield resources [ 599.980602] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.980602] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] self.driver.spawn(context, instance, image_meta, [ 599.980602] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 599.980602] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.980602] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.980602] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] vm_ref = self.build_virtual_machine(instance, [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] for vif in network_info: [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] return self._sync_wrapper(fn, *args, **kwargs) [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] self.wait() [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] self[:] = self._gt.wait() [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] return self._exit_event.wait() [ 599.981046] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] result = hub.switch() [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] return self.greenlet.switch() [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] result = function(*args, **kwargs) [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] return func(*args, **kwargs) [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] raise e [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] nwinfo = self.network_api.allocate_for_instance( [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.981526] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] created_port_ids = self._update_ports_for_instance( [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] with excutils.save_and_reraise_exception(): [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] self.force_reraise() [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] raise self.value [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] updated_port = self._update_port( [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] _ensure_no_port_binding_failure(port) [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.981994] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] raise exception.PortBindingFailed(port_id=port['id']) [ 599.982514] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] nova.exception.PortBindingFailed: Binding failed for port c54cc217-14c7-43a7-9f9c-a980c39413ee, please check neutron logs for more information. [ 599.982514] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] [ 599.982514] env[63293]: INFO nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Terminating instance [ 599.982514] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Acquiring lock "refresh_cache-25f0de66-914f-46b9-a15b-d936df5b3ab4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.982514] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Acquired lock "refresh_cache-25f0de66-914f-46b9-a15b-d936df5b3ab4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.982514] env[63293]: DEBUG nova.network.neutron [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 600.426142] env[63293]: DEBUG oslo_concurrency.lockutils [None req-48e6ba62-1c09-45c4-8828-daa8173056f3 tempest-ServersAdmin275Test-686916295 tempest-ServersAdmin275Test-686916295-project-member] Lock "e80209b3-273c-4045-ad8a-2a3c4c48200a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.538s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.442281] env[63293]: INFO nova.compute.manager [-] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Took 1.03 seconds to deallocate network for instance. [ 600.444320] env[63293]: DEBUG nova.compute.claims [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.444494] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.509090] env[63293]: DEBUG nova.network.neutron [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.608683] env[63293]: DEBUG nova.network.neutron [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.111165] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Releasing lock "refresh_cache-25f0de66-914f-46b9-a15b-d936df5b3ab4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.111582] env[63293]: DEBUG nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 601.111768] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 601.112068] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30d37bdd-346e-438e-95b1-253504d41d10 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.122755] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71bcb081-6801-4422-b984-58e4d14ac7fe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.146670] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 25f0de66-914f-46b9-a15b-d936df5b3ab4 could not be found. [ 601.146670] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 601.146785] env[63293]: INFO nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 601.147610] env[63293]: DEBUG oslo.service.loopingcall [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.149661] env[63293]: DEBUG nova.compute.manager [-] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.149759] env[63293]: DEBUG nova.network.neutron [-] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.164152] env[63293]: DEBUG nova.network.neutron [-] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.229173] env[63293]: DEBUG nova.compute.manager [req-dff8f1fe-4f55-4d27-9128-e893f6dccf62 req-40e84b9b-8a0e-4c99-a88c-a0f8e8af7b5b service nova] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Received event network-changed-c54cc217-14c7-43a7-9f9c-a980c39413ee {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.229409] env[63293]: DEBUG nova.compute.manager [req-dff8f1fe-4f55-4d27-9128-e893f6dccf62 req-40e84b9b-8a0e-4c99-a88c-a0f8e8af7b5b service nova] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Refreshing instance network info cache due to event network-changed-c54cc217-14c7-43a7-9f9c-a980c39413ee. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 601.229632] env[63293]: DEBUG oslo_concurrency.lockutils [req-dff8f1fe-4f55-4d27-9128-e893f6dccf62 req-40e84b9b-8a0e-4c99-a88c-a0f8e8af7b5b service nova] Acquiring lock "refresh_cache-25f0de66-914f-46b9-a15b-d936df5b3ab4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.229758] env[63293]: DEBUG oslo_concurrency.lockutils [req-dff8f1fe-4f55-4d27-9128-e893f6dccf62 req-40e84b9b-8a0e-4c99-a88c-a0f8e8af7b5b service nova] Acquired lock "refresh_cache-25f0de66-914f-46b9-a15b-d936df5b3ab4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.229913] env[63293]: DEBUG nova.network.neutron [req-dff8f1fe-4f55-4d27-9128-e893f6dccf62 req-40e84b9b-8a0e-4c99-a88c-a0f8e8af7b5b service nova] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Refreshing network info cache for port c54cc217-14c7-43a7-9f9c-a980c39413ee {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 601.311084] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62ae904-4e01-4943-9822-3da645a2879b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.320301] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e34e6a-4fa9-46e9-a559-e645aea12657 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.348750] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569a74fd-233a-49ef-8845-0ffb9104b021 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.356466] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bac1417-3587-4819-80a1-026fe8b3b637 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.371627] env[63293]: DEBUG nova.compute.provider_tree [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.666586] env[63293]: DEBUG nova.network.neutron [-] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.764549] env[63293]: DEBUG nova.network.neutron [req-dff8f1fe-4f55-4d27-9128-e893f6dccf62 req-40e84b9b-8a0e-4c99-a88c-a0f8e8af7b5b service nova] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.866326] env[63293]: DEBUG nova.network.neutron [req-dff8f1fe-4f55-4d27-9128-e893f6dccf62 req-40e84b9b-8a0e-4c99-a88c-a0f8e8af7b5b service nova] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.875583] env[63293]: DEBUG nova.scheduler.client.report [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.170152] env[63293]: INFO nova.compute.manager [-] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Took 1.02 seconds to deallocate network for instance. [ 602.171966] env[63293]: DEBUG nova.compute.claims [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 602.172181] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.369780] env[63293]: DEBUG oslo_concurrency.lockutils [req-dff8f1fe-4f55-4d27-9128-e893f6dccf62 req-40e84b9b-8a0e-4c99-a88c-a0f8e8af7b5b service nova] Releasing lock "refresh_cache-25f0de66-914f-46b9-a15b-d936df5b3ab4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.370055] env[63293]: DEBUG nova.compute.manager [req-dff8f1fe-4f55-4d27-9128-e893f6dccf62 req-40e84b9b-8a0e-4c99-a88c-a0f8e8af7b5b service nova] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Received event network-vif-deleted-c54cc217-14c7-43a7-9f9c-a980c39413ee {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.380126] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.380624] env[63293]: DEBUG nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.383240] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.898s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.385233] env[63293]: INFO nova.compute.claims [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.890199] env[63293]: DEBUG nova.compute.utils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.894173] env[63293]: DEBUG nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.894412] env[63293]: DEBUG nova.network.neutron [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 602.963519] env[63293]: DEBUG nova.policy [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd8e80cd3dca42989849fbfbeec2adf3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cc2129595c9a4f5793e1f35a760b1393', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 603.395035] env[63293]: DEBUG nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.475010] env[63293]: DEBUG nova.network.neutron [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Successfully created port: ddc895f1-d2a2-4a78-8e01-0c6b6660dee7 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.925693] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d576ac3-57b7-4a33-af86-65c531a642c4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.936017] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbcfeb5a-3090-4069-bcf6-a889d20a2205 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.970743] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02699fce-f9ec-4706-9a43-456e0d32bb2c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.979032] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2564e70-6682-4d25-a2ac-98bdb3ba51c8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.995436] env[63293]: DEBUG nova.compute.provider_tree [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.411381] env[63293]: DEBUG nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.445362] env[63293]: DEBUG nova.virt.hardware [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.445635] env[63293]: DEBUG nova.virt.hardware [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.445813] env[63293]: DEBUG nova.virt.hardware [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.446461] env[63293]: DEBUG nova.virt.hardware [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.446461] env[63293]: DEBUG nova.virt.hardware [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.446461] env[63293]: DEBUG nova.virt.hardware [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.446588] env[63293]: DEBUG nova.virt.hardware [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.446763] env[63293]: DEBUG nova.virt.hardware [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.446992] env[63293]: DEBUG nova.virt.hardware [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.447249] env[63293]: DEBUG nova.virt.hardware [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.447455] env[63293]: DEBUG nova.virt.hardware [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.448389] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80166c0-950a-471d-bb68-39594dfffce6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.456559] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7070484-eef9-4759-b391-98df6bcba601 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.499170] env[63293]: DEBUG nova.scheduler.client.report [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.527141] env[63293]: DEBUG nova.compute.manager [req-84862e04-369e-4c37-afe7-8a4b79298344 req-d5c5c10b-f84a-4709-bfbd-31fa83a22602 service nova] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Received event network-changed-ddc895f1-d2a2-4a78-8e01-0c6b6660dee7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.527141] env[63293]: DEBUG nova.compute.manager [req-84862e04-369e-4c37-afe7-8a4b79298344 req-d5c5c10b-f84a-4709-bfbd-31fa83a22602 service nova] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Refreshing instance network info cache due to event network-changed-ddc895f1-d2a2-4a78-8e01-0c6b6660dee7. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 604.527248] env[63293]: DEBUG oslo_concurrency.lockutils [req-84862e04-369e-4c37-afe7-8a4b79298344 req-d5c5c10b-f84a-4709-bfbd-31fa83a22602 service nova] Acquiring lock "refresh_cache-c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.527351] env[63293]: DEBUG oslo_concurrency.lockutils [req-84862e04-369e-4c37-afe7-8a4b79298344 req-d5c5c10b-f84a-4709-bfbd-31fa83a22602 service nova] Acquired lock "refresh_cache-c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.527506] env[63293]: DEBUG nova.network.neutron [req-84862e04-369e-4c37-afe7-8a4b79298344 req-d5c5c10b-f84a-4709-bfbd-31fa83a22602 service nova] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Refreshing network info cache for port ddc895f1-d2a2-4a78-8e01-0c6b6660dee7 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 604.755019] env[63293]: ERROR nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ddc895f1-d2a2-4a78-8e01-0c6b6660dee7, please check neutron logs for more information. [ 604.755019] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 604.755019] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.755019] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 604.755019] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.755019] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 604.755019] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.755019] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 604.755019] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.755019] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 604.755019] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.755019] env[63293]: ERROR nova.compute.manager raise self.value [ 604.755019] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.755019] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 604.755019] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.755019] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 604.755552] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.755552] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 604.755552] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ddc895f1-d2a2-4a78-8e01-0c6b6660dee7, please check neutron logs for more information. [ 604.755552] env[63293]: ERROR nova.compute.manager [ 604.755552] env[63293]: Traceback (most recent call last): [ 604.755552] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 604.755552] env[63293]: listener.cb(fileno) [ 604.755552] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.755552] env[63293]: result = function(*args, **kwargs) [ 604.755552] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.755552] env[63293]: return func(*args, **kwargs) [ 604.755552] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.755552] env[63293]: raise e [ 604.755552] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.755552] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 604.755552] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.755552] env[63293]: created_port_ids = self._update_ports_for_instance( [ 604.755552] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.755552] env[63293]: with excutils.save_and_reraise_exception(): [ 604.755552] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.755552] env[63293]: self.force_reraise() [ 604.755552] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.755552] env[63293]: raise self.value [ 604.755552] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.755552] env[63293]: updated_port = self._update_port( [ 604.755552] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.755552] env[63293]: _ensure_no_port_binding_failure(port) [ 604.755552] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.755552] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 604.756262] env[63293]: nova.exception.PortBindingFailed: Binding failed for port ddc895f1-d2a2-4a78-8e01-0c6b6660dee7, please check neutron logs for more information. [ 604.756262] env[63293]: Removing descriptor: 18 [ 604.756262] env[63293]: ERROR nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ddc895f1-d2a2-4a78-8e01-0c6b6660dee7, please check neutron logs for more information. [ 604.756262] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Traceback (most recent call last): [ 604.756262] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 604.756262] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] yield resources [ 604.756262] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.756262] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] self.driver.spawn(context, instance, image_meta, [ 604.756262] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 604.756262] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.756262] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.756262] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] vm_ref = self.build_virtual_machine(instance, [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] for vif in network_info: [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] return self._sync_wrapper(fn, *args, **kwargs) [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] self.wait() [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] self[:] = self._gt.wait() [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] return self._exit_event.wait() [ 604.756562] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] result = hub.switch() [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] return self.greenlet.switch() [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] result = function(*args, **kwargs) [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] return func(*args, **kwargs) [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] raise e [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] nwinfo = self.network_api.allocate_for_instance( [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.756852] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] created_port_ids = self._update_ports_for_instance( [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] with excutils.save_and_reraise_exception(): [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] self.force_reraise() [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] raise self.value [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] updated_port = self._update_port( [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] _ensure_no_port_binding_failure(port) [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.757212] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] raise exception.PortBindingFailed(port_id=port['id']) [ 604.757507] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] nova.exception.PortBindingFailed: Binding failed for port ddc895f1-d2a2-4a78-8e01-0c6b6660dee7, please check neutron logs for more information. [ 604.757507] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] [ 604.757507] env[63293]: INFO nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Terminating instance [ 604.757507] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Acquiring lock "refresh_cache-c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.004122] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.004605] env[63293]: DEBUG nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 605.007510] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.104s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.056174] env[63293]: DEBUG nova.network.neutron [req-84862e04-369e-4c37-afe7-8a4b79298344 req-d5c5c10b-f84a-4709-bfbd-31fa83a22602 service nova] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.141538] env[63293]: DEBUG nova.network.neutron [req-84862e04-369e-4c37-afe7-8a4b79298344 req-d5c5c10b-f84a-4709-bfbd-31fa83a22602 service nova] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.512110] env[63293]: DEBUG nova.compute.utils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.516161] env[63293]: DEBUG nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 605.516414] env[63293]: DEBUG nova.network.neutron [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 605.572259] env[63293]: DEBUG nova.policy [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a141691de4b442a392e083ad1423ecaa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e84f511614aa4e06915c25edbb3a9ce7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 605.644352] env[63293]: DEBUG oslo_concurrency.lockutils [req-84862e04-369e-4c37-afe7-8a4b79298344 req-d5c5c10b-f84a-4709-bfbd-31fa83a22602 service nova] Releasing lock "refresh_cache-c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.645180] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Acquired lock "refresh_cache-c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.645180] env[63293]: DEBUG nova.network.neutron [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.902693] env[63293]: DEBUG nova.network.neutron [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Successfully created port: bfb28db2-e515-4222-9f85-aace51153a77 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 605.930851] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3853816c-f04e-4706-a3cd-7b5ece2df304 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.939669] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa966802-7b6a-4536-8d96-b5e8506f0c22 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.974361] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904abea6-f5a4-475c-bbcb-ee80010daf7e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.982248] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbbd33e-50b5-4290-80a4-dcdb4b12f7d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.000079] env[63293]: DEBUG nova.compute.provider_tree [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.005034] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Acquiring lock "263d1329-a81a-4d50-91b6-3927bffe8d4c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.006424] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Lock "263d1329-a81a-4d50-91b6-3927bffe8d4c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.018478] env[63293]: DEBUG nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 606.199112] env[63293]: DEBUG nova.network.neutron [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.378347] env[63293]: DEBUG nova.network.neutron [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.506433] env[63293]: DEBUG nova.scheduler.client.report [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.582301] env[63293]: DEBUG nova.compute.manager [req-356ef97a-2150-41e2-8f74-f05398fdd216 req-a0544f8a-ade3-43b2-8427-c0c1900b8eac service nova] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Received event network-vif-deleted-ddc895f1-d2a2-4a78-8e01-0c6b6660dee7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 606.883246] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Releasing lock "refresh_cache-c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.883669] env[63293]: DEBUG nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 606.887071] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 606.887071] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37f17e15-72f5-4994-9454-6a768b64a502 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.893283] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dddca1b-0885-40da-a7f6-525c46e74f60 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.914501] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c1f9f075-7d1c-4c05-9dfd-43820d2d07ad could not be found. [ 606.915021] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 606.915021] env[63293]: INFO nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Took 0.03 seconds to destroy the instance on the hypervisor. [ 606.915127] env[63293]: DEBUG oslo.service.loopingcall [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.915331] env[63293]: DEBUG nova.compute.manager [-] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.915420] env[63293]: DEBUG nova.network.neutron [-] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.934445] env[63293]: DEBUG nova.network.neutron [-] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.012026] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.012892] env[63293]: ERROR nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cb3b734d-a172-403d-a199-800fe83a10ea, please check neutron logs for more information. [ 607.012892] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Traceback (most recent call last): [ 607.012892] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.012892] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] self.driver.spawn(context, instance, image_meta, [ 607.012892] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 607.012892] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.012892] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.012892] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] vm_ref = self.build_virtual_machine(instance, [ 607.012892] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.012892] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.012892] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] for vif in network_info: [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] return self._sync_wrapper(fn, *args, **kwargs) [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] self.wait() [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] self[:] = self._gt.wait() [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] return self._exit_event.wait() [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] result = hub.switch() [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.013232] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] return self.greenlet.switch() [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] result = function(*args, **kwargs) [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] return func(*args, **kwargs) [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] raise e [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] nwinfo = self.network_api.allocate_for_instance( [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] created_port_ids = self._update_ports_for_instance( [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] with excutils.save_and_reraise_exception(): [ 607.013562] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.013934] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] self.force_reraise() [ 607.013934] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.013934] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] raise self.value [ 607.013934] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.013934] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] updated_port = self._update_port( [ 607.013934] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.013934] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] _ensure_no_port_binding_failure(port) [ 607.013934] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.013934] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] raise exception.PortBindingFailed(port_id=port['id']) [ 607.013934] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] nova.exception.PortBindingFailed: Binding failed for port cb3b734d-a172-403d-a199-800fe83a10ea, please check neutron logs for more information. [ 607.013934] env[63293]: ERROR nova.compute.manager [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] [ 607.014199] env[63293]: DEBUG nova.compute.utils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Binding failed for port cb3b734d-a172-403d-a199-800fe83a10ea, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 607.014571] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.660s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.018600] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Build of instance 43582f6d-75fb-4734-9fb0-b71db3ddf74c was re-scheduled: Binding failed for port cb3b734d-a172-403d-a199-800fe83a10ea, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 607.019104] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 607.022519] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "refresh_cache-43582f6d-75fb-4734-9fb0-b71db3ddf74c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.022519] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquired lock "refresh_cache-43582f6d-75fb-4734-9fb0-b71db3ddf74c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.022519] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 607.028305] env[63293]: DEBUG nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 607.063526] env[63293]: DEBUG nova.virt.hardware [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 607.063774] env[63293]: DEBUG nova.virt.hardware [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 607.063928] env[63293]: DEBUG nova.virt.hardware [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 607.064127] env[63293]: DEBUG nova.virt.hardware [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 607.064277] env[63293]: DEBUG nova.virt.hardware [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 607.064422] env[63293]: DEBUG nova.virt.hardware [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 607.064625] env[63293]: DEBUG nova.virt.hardware [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 607.064785] env[63293]: DEBUG nova.virt.hardware [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 607.064948] env[63293]: DEBUG nova.virt.hardware [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 607.065149] env[63293]: DEBUG nova.virt.hardware [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 607.065333] env[63293]: DEBUG nova.virt.hardware [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 607.066910] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21455e5-0153-49eb-bd6e-0c13d55c8431 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.077447] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e526c8-545e-440a-988c-9dd3fb82e5c8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.179324] env[63293]: ERROR nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bfb28db2-e515-4222-9f85-aace51153a77, please check neutron logs for more information. [ 607.179324] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 607.179324] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.179324] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 607.179324] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.179324] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 607.179324] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.179324] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 607.179324] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.179324] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 607.179324] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.179324] env[63293]: ERROR nova.compute.manager raise self.value [ 607.179324] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.179324] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 607.179324] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.179324] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 607.179698] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.179698] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 607.179698] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bfb28db2-e515-4222-9f85-aace51153a77, please check neutron logs for more information. [ 607.179698] env[63293]: ERROR nova.compute.manager [ 607.179698] env[63293]: Traceback (most recent call last): [ 607.179698] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 607.179698] env[63293]: listener.cb(fileno) [ 607.179698] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.179698] env[63293]: result = function(*args, **kwargs) [ 607.179698] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 607.179698] env[63293]: return func(*args, **kwargs) [ 607.179698] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.179698] env[63293]: raise e [ 607.179698] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.179698] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 607.179698] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.179698] env[63293]: created_port_ids = self._update_ports_for_instance( [ 607.179698] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.179698] env[63293]: with excutils.save_and_reraise_exception(): [ 607.179698] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.179698] env[63293]: self.force_reraise() [ 607.179698] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.179698] env[63293]: raise self.value [ 607.179698] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.179698] env[63293]: updated_port = self._update_port( [ 607.179698] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.179698] env[63293]: _ensure_no_port_binding_failure(port) [ 607.179698] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.179698] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 607.180344] env[63293]: nova.exception.PortBindingFailed: Binding failed for port bfb28db2-e515-4222-9f85-aace51153a77, please check neutron logs for more information. [ 607.180344] env[63293]: Removing descriptor: 18 [ 607.180344] env[63293]: ERROR nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bfb28db2-e515-4222-9f85-aace51153a77, please check neutron logs for more information. [ 607.180344] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Traceback (most recent call last): [ 607.180344] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 607.180344] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] yield resources [ 607.180344] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.180344] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] self.driver.spawn(context, instance, image_meta, [ 607.180344] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 607.180344] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.180344] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.180344] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] vm_ref = self.build_virtual_machine(instance, [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] for vif in network_info: [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] return self._sync_wrapper(fn, *args, **kwargs) [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] self.wait() [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] self[:] = self._gt.wait() [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] return self._exit_event.wait() [ 607.180599] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] result = hub.switch() [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] return self.greenlet.switch() [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] result = function(*args, **kwargs) [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] return func(*args, **kwargs) [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] raise e [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] nwinfo = self.network_api.allocate_for_instance( [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.180869] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] created_port_ids = self._update_ports_for_instance( [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] with excutils.save_and_reraise_exception(): [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] self.force_reraise() [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] raise self.value [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] updated_port = self._update_port( [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] _ensure_no_port_binding_failure(port) [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.181143] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] raise exception.PortBindingFailed(port_id=port['id']) [ 607.181388] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] nova.exception.PortBindingFailed: Binding failed for port bfb28db2-e515-4222-9f85-aace51153a77, please check neutron logs for more information. [ 607.181388] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] [ 607.181388] env[63293]: INFO nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Terminating instance [ 607.182479] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Acquiring lock "refresh_cache-a236be17-8965-4785-834e-9ce22eeb75eb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.182632] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Acquired lock "refresh_cache-a236be17-8965-4785-834e-9ce22eeb75eb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.182794] env[63293]: DEBUG nova.network.neutron [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 607.436970] env[63293]: DEBUG nova.network.neutron [-] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.541178] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.651488] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.696825] env[63293]: DEBUG nova.network.neutron [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.778701] env[63293]: DEBUG nova.network.neutron [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.885712] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86e0533-670b-4175-a520-7d91a723beb6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.893439] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1927551c-dad5-4bb4-8d5c-8cdc82c14270 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.922636] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a5489c-b3a1-4859-b63f-4a119314db80 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.929442] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e127a44-3061-4ee0-a17c-cbcabbe1770b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.948863] env[63293]: INFO nova.compute.manager [-] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Took 1.03 seconds to deallocate network for instance. [ 607.949454] env[63293]: DEBUG nova.compute.provider_tree [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.953662] env[63293]: DEBUG nova.compute.claims [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 607.953662] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.153866] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Releasing lock "refresh_cache-43582f6d-75fb-4734-9fb0-b71db3ddf74c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.154260] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 608.154526] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.154728] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 608.172023] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.283665] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Releasing lock "refresh_cache-a236be17-8965-4785-834e-9ce22eeb75eb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.283833] env[63293]: DEBUG nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.284018] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 608.284372] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-413cab6d-f405-4d9f-b110-e591bae51f12 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.293408] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3089ff-ad97-4fdd-b68a-0b5923db749e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.314559] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a236be17-8965-4785-834e-9ce22eeb75eb could not be found. [ 608.314913] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 608.314984] env[63293]: INFO nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 608.315194] env[63293]: DEBUG oslo.service.loopingcall [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.315423] env[63293]: DEBUG nova.compute.manager [-] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.315508] env[63293]: DEBUG nova.network.neutron [-] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 608.330944] env[63293]: DEBUG nova.network.neutron [-] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.454564] env[63293]: DEBUG nova.scheduler.client.report [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.616225] env[63293]: DEBUG nova.compute.manager [req-d3af082b-cacf-4f8f-b4db-e5a610a0498f req-7f535884-3f07-4ca6-93d1-c162c5028b40 service nova] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Received event network-changed-bfb28db2-e515-4222-9f85-aace51153a77 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.616225] env[63293]: DEBUG nova.compute.manager [req-d3af082b-cacf-4f8f-b4db-e5a610a0498f req-7f535884-3f07-4ca6-93d1-c162c5028b40 service nova] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Refreshing instance network info cache due to event network-changed-bfb28db2-e515-4222-9f85-aace51153a77. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 608.616225] env[63293]: DEBUG oslo_concurrency.lockutils [req-d3af082b-cacf-4f8f-b4db-e5a610a0498f req-7f535884-3f07-4ca6-93d1-c162c5028b40 service nova] Acquiring lock "refresh_cache-a236be17-8965-4785-834e-9ce22eeb75eb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.616225] env[63293]: DEBUG oslo_concurrency.lockutils [req-d3af082b-cacf-4f8f-b4db-e5a610a0498f req-7f535884-3f07-4ca6-93d1-c162c5028b40 service nova] Acquired lock "refresh_cache-a236be17-8965-4785-834e-9ce22eeb75eb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.616225] env[63293]: DEBUG nova.network.neutron [req-d3af082b-cacf-4f8f-b4db-e5a610a0498f req-7f535884-3f07-4ca6-93d1-c162c5028b40 service nova] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Refreshing network info cache for port bfb28db2-e515-4222-9f85-aace51153a77 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 608.674599] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.833234] env[63293]: DEBUG nova.network.neutron [-] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.960758] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.945s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.960758] env[63293]: ERROR nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 60ec81ef-8266-4005-aef5-f24d13bcf605, please check neutron logs for more information. [ 608.960758] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Traceback (most recent call last): [ 608.960758] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.960758] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] self.driver.spawn(context, instance, image_meta, [ 608.960758] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 608.960758] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.960758] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.960758] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] vm_ref = self.build_virtual_machine(instance, [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] for vif in network_info: [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] return self._sync_wrapper(fn, *args, **kwargs) [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] self.wait() [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] self[:] = self._gt.wait() [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] return self._exit_event.wait() [ 608.961136] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] result = hub.switch() [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] return self.greenlet.switch() [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] result = function(*args, **kwargs) [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] return func(*args, **kwargs) [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] raise e [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] nwinfo = self.network_api.allocate_for_instance( [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.961472] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] created_port_ids = self._update_ports_for_instance( [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] with excutils.save_and_reraise_exception(): [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] self.force_reraise() [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] raise self.value [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] updated_port = self._update_port( [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] _ensure_no_port_binding_failure(port) [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.961842] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] raise exception.PortBindingFailed(port_id=port['id']) [ 608.962112] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] nova.exception.PortBindingFailed: Binding failed for port 60ec81ef-8266-4005-aef5-f24d13bcf605, please check neutron logs for more information. [ 608.962112] env[63293]: ERROR nova.compute.manager [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] [ 608.962112] env[63293]: DEBUG nova.compute.utils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Binding failed for port 60ec81ef-8266-4005-aef5-f24d13bcf605, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 608.962112] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.879s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.965035] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Build of instance 0b381f4c-372f-41c8-9661-7e6bab6dd306 was re-scheduled: Binding failed for port 60ec81ef-8266-4005-aef5-f24d13bcf605, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 608.965470] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 608.965692] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquiring lock "refresh_cache-0b381f4c-372f-41c8-9661-7e6bab6dd306" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.965838] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Acquired lock "refresh_cache-0b381f4c-372f-41c8-9661-7e6bab6dd306" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.965993] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 609.136883] env[63293]: DEBUG nova.network.neutron [req-d3af082b-cacf-4f8f-b4db-e5a610a0498f req-7f535884-3f07-4ca6-93d1-c162c5028b40 service nova] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.177528] env[63293]: INFO nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 43582f6d-75fb-4734-9fb0-b71db3ddf74c] Took 1.02 seconds to deallocate network for instance. [ 609.239719] env[63293]: DEBUG nova.network.neutron [req-d3af082b-cacf-4f8f-b4db-e5a610a0498f req-7f535884-3f07-4ca6-93d1-c162c5028b40 service nova] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.336533] env[63293]: INFO nova.compute.manager [-] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Took 1.02 seconds to deallocate network for instance. [ 609.339143] env[63293]: DEBUG nova.compute.claims [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 609.339314] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.484294] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.580665] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.742903] env[63293]: DEBUG oslo_concurrency.lockutils [req-d3af082b-cacf-4f8f-b4db-e5a610a0498f req-7f535884-3f07-4ca6-93d1-c162c5028b40 service nova] Releasing lock "refresh_cache-a236be17-8965-4785-834e-9ce22eeb75eb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.743168] env[63293]: DEBUG nova.compute.manager [req-d3af082b-cacf-4f8f-b4db-e5a610a0498f req-7f535884-3f07-4ca6-93d1-c162c5028b40 service nova] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Received event network-vif-deleted-bfb28db2-e515-4222-9f85-aace51153a77 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 609.842922] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92c5eb7-6c5b-49b1-bd12-2b9a1b030e29 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.852132] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a49631f-624b-4acd-95bd-29759d5730e5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.883058] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a03371e-5f1c-4277-8e12-80fb0972378c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.889346] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743e3147-312e-4be6-9815-753e632f8aca {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.902219] env[63293]: DEBUG nova.compute.provider_tree [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.083149] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Releasing lock "refresh_cache-0b381f4c-372f-41c8-9661-7e6bab6dd306" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.083325] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 610.083515] env[63293]: DEBUG nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.083683] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 610.097449] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.204230] env[63293]: INFO nova.scheduler.client.report [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Deleted allocations for instance 43582f6d-75fb-4734-9fb0-b71db3ddf74c [ 610.406647] env[63293]: DEBUG nova.scheduler.client.report [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.601907] env[63293]: DEBUG nova.network.neutron [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.712928] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "43582f6d-75fb-4734-9fb0-b71db3ddf74c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.891s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.911040] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.949s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.911606] env[63293]: ERROR nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c, please check neutron logs for more information. [ 610.911606] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Traceback (most recent call last): [ 610.911606] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.911606] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] self.driver.spawn(context, instance, image_meta, [ 610.911606] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.911606] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.911606] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.911606] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] vm_ref = self.build_virtual_machine(instance, [ 610.911606] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.911606] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.911606] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] for vif in network_info: [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] return self._sync_wrapper(fn, *args, **kwargs) [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] self.wait() [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] self[:] = self._gt.wait() [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] return self._exit_event.wait() [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] result = hub.switch() [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.911877] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] return self.greenlet.switch() [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] result = function(*args, **kwargs) [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] return func(*args, **kwargs) [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] raise e [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] nwinfo = self.network_api.allocate_for_instance( [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] created_port_ids = self._update_ports_for_instance( [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] with excutils.save_and_reraise_exception(): [ 610.912314] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.912802] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] self.force_reraise() [ 610.912802] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.912802] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] raise self.value [ 610.912802] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.912802] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] updated_port = self._update_port( [ 610.912802] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.912802] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] _ensure_no_port_binding_failure(port) [ 610.912802] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.912802] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] raise exception.PortBindingFailed(port_id=port['id']) [ 610.912802] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] nova.exception.PortBindingFailed: Binding failed for port 0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c, please check neutron logs for more information. [ 610.912802] env[63293]: ERROR nova.compute.manager [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] [ 610.913248] env[63293]: DEBUG nova.compute.utils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Binding failed for port 0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 610.913538] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.415s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.916314] env[63293]: DEBUG nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Build of instance 4dce2ec6-6f7b-4294-a32e-4cd222ae3492 was re-scheduled: Binding failed for port 0d7bd4b1-cdbf-4282-8c18-1a6ddcbd8f0c, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 610.916730] env[63293]: DEBUG nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 610.916964] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-4dce2ec6-6f7b-4294-a32e-4cd222ae3492" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.917154] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-4dce2ec6-6f7b-4294-a32e-4cd222ae3492" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.917305] env[63293]: DEBUG nova.network.neutron [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.104574] env[63293]: INFO nova.compute.manager [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] [instance: 0b381f4c-372f-41c8-9661-7e6bab6dd306] Took 1.02 seconds to deallocate network for instance. [ 611.215456] env[63293]: DEBUG nova.compute.manager [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 611.442716] env[63293]: DEBUG nova.network.neutron [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.532375] env[63293]: DEBUG nova.network.neutron [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.731284] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.778690] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0062b1-91a5-454b-89cb-ba411d7a3252 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.786230] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cfa511a-239d-4efe-9661-a19525dd4fd2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.815435] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9752659e-fa91-442e-9d1f-ef69223a8461 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.822820] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a0175e3-dead-48ad-8250-73601e24c3a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.835679] env[63293]: DEBUG nova.compute.provider_tree [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.035679] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-4dce2ec6-6f7b-4294-a32e-4cd222ae3492" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.035934] env[63293]: DEBUG nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 612.036069] env[63293]: DEBUG nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.036240] env[63293]: DEBUG nova.network.neutron [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.050694] env[63293]: DEBUG nova.network.neutron [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.132264] env[63293]: INFO nova.scheduler.client.report [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Deleted allocations for instance 0b381f4c-372f-41c8-9661-7e6bab6dd306 [ 612.339307] env[63293]: DEBUG nova.scheduler.client.report [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.553072] env[63293]: DEBUG nova.network.neutron [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.644093] env[63293]: DEBUG oslo_concurrency.lockutils [None req-88a58596-da6a-4c17-b422-d3f1defe5643 tempest-ListServersNegativeTestJSON-1022782966 tempest-ListServersNegativeTestJSON-1022782966-project-member] Lock "0b381f4c-372f-41c8-9661-7e6bab6dd306" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.756s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.843719] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.930s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.844376] env[63293]: ERROR nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9567745f-7393-4661-994a-fb9813f6b138, please check neutron logs for more information. [ 612.844376] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Traceback (most recent call last): [ 612.844376] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 612.844376] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] self.driver.spawn(context, instance, image_meta, [ 612.844376] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 612.844376] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.844376] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.844376] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] vm_ref = self.build_virtual_machine(instance, [ 612.844376] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.844376] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.844376] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] for vif in network_info: [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] return self._sync_wrapper(fn, *args, **kwargs) [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] self.wait() [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] self[:] = self._gt.wait() [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] return self._exit_event.wait() [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] current.throw(*self._exc) [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.844702] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] result = function(*args, **kwargs) [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] return func(*args, **kwargs) [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] raise e [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] nwinfo = self.network_api.allocate_for_instance( [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] created_port_ids = self._update_ports_for_instance( [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] with excutils.save_and_reraise_exception(): [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] self.force_reraise() [ 612.845110] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.845437] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] raise self.value [ 612.845437] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.845437] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] updated_port = self._update_port( [ 612.845437] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.845437] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] _ensure_no_port_binding_failure(port) [ 612.845437] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.845437] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] raise exception.PortBindingFailed(port_id=port['id']) [ 612.845437] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] nova.exception.PortBindingFailed: Binding failed for port 9567745f-7393-4661-994a-fb9813f6b138, please check neutron logs for more information. [ 612.845437] env[63293]: ERROR nova.compute.manager [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] [ 612.845437] env[63293]: DEBUG nova.compute.utils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Binding failed for port 9567745f-7393-4661-994a-fb9813f6b138, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.846325] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.671s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.849148] env[63293]: DEBUG nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Build of instance 1ac2f52a-27ef-4ce4-9540-288658f08344 was re-scheduled: Binding failed for port 9567745f-7393-4661-994a-fb9813f6b138, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 612.849580] env[63293]: DEBUG nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 612.849808] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquiring lock "refresh_cache-1ac2f52a-27ef-4ce4-9540-288658f08344" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.849955] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Acquired lock "refresh_cache-1ac2f52a-27ef-4ce4-9540-288658f08344" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.850124] env[63293]: DEBUG nova.network.neutron [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.056977] env[63293]: INFO nova.compute.manager [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 4dce2ec6-6f7b-4294-a32e-4cd222ae3492] Took 1.02 seconds to deallocate network for instance. [ 613.146108] env[63293]: DEBUG nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 613.372061] env[63293]: DEBUG nova.network.neutron [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.462835] env[63293]: DEBUG nova.network.neutron [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.664946] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.762474] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18691227-9916-40cb-a295-a94f1c6201eb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.770082] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d041c0-617b-4f18-8703-656c0bbc7bc7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.802150] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eccaa940-5d3f-4748-b35b-432f2dc6ba76 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.809751] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e22efe-8c14-4acc-b44f-8e33531bf4f9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.823230] env[63293]: DEBUG nova.compute.provider_tree [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.964825] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Releasing lock "refresh_cache-1ac2f52a-27ef-4ce4-9540-288658f08344" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.964960] env[63293]: DEBUG nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 613.965174] env[63293]: DEBUG nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.965345] env[63293]: DEBUG nova.network.neutron [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.983907] env[63293]: DEBUG nova.network.neutron [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.091923] env[63293]: INFO nova.scheduler.client.report [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleted allocations for instance 4dce2ec6-6f7b-4294-a32e-4cd222ae3492 [ 614.326065] env[63293]: DEBUG nova.scheduler.client.report [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.487760] env[63293]: DEBUG nova.network.neutron [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.603356] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5091b5a-9197-49b0-9dc1-68dc15813603 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "4dce2ec6-6f7b-4294-a32e-4cd222ae3492" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.598s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.831476] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.985s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.832181] env[63293]: ERROR nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d6936f7c-3207-4cca-92d7-bdc13912b1ab, please check neutron logs for more information. [ 614.832181] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] Traceback (most recent call last): [ 614.832181] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.832181] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] self.driver.spawn(context, instance, image_meta, [ 614.832181] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 614.832181] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.832181] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.832181] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] vm_ref = self.build_virtual_machine(instance, [ 614.832181] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.832181] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.832181] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] for vif in network_info: [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] return self._sync_wrapper(fn, *args, **kwargs) [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] self.wait() [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] self[:] = self._gt.wait() [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] return self._exit_event.wait() [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] result = hub.switch() [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.832656] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] return self.greenlet.switch() [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] result = function(*args, **kwargs) [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] return func(*args, **kwargs) [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] raise e [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] nwinfo = self.network_api.allocate_for_instance( [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] created_port_ids = self._update_ports_for_instance( [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] with excutils.save_and_reraise_exception(): [ 614.833561] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.833994] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] self.force_reraise() [ 614.833994] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.833994] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] raise self.value [ 614.833994] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.833994] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] updated_port = self._update_port( [ 614.833994] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.833994] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] _ensure_no_port_binding_failure(port) [ 614.833994] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.833994] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] raise exception.PortBindingFailed(port_id=port['id']) [ 614.833994] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] nova.exception.PortBindingFailed: Binding failed for port d6936f7c-3207-4cca-92d7-bdc13912b1ab, please check neutron logs for more information. [ 614.833994] env[63293]: ERROR nova.compute.manager [instance: db725426-1b67-437c-aea5-cf0da50454a2] [ 614.834376] env[63293]: DEBUG nova.compute.utils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Binding failed for port d6936f7c-3207-4cca-92d7-bdc13912b1ab, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 614.834376] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.375s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.835545] env[63293]: INFO nova.compute.claims [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.838366] env[63293]: DEBUG nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Build of instance db725426-1b67-437c-aea5-cf0da50454a2 was re-scheduled: Binding failed for port d6936f7c-3207-4cca-92d7-bdc13912b1ab, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 614.838788] env[63293]: DEBUG nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 614.839028] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquiring lock "refresh_cache-db725426-1b67-437c-aea5-cf0da50454a2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.839181] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Acquired lock "refresh_cache-db725426-1b67-437c-aea5-cf0da50454a2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.839341] env[63293]: DEBUG nova.network.neutron [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 614.991755] env[63293]: INFO nova.compute.manager [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] [instance: 1ac2f52a-27ef-4ce4-9540-288658f08344] Took 1.03 seconds to deallocate network for instance. [ 615.105963] env[63293]: DEBUG nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 615.362964] env[63293]: DEBUG nova.network.neutron [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.471827] env[63293]: DEBUG nova.network.neutron [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.638794] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.976073] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Releasing lock "refresh_cache-db725426-1b67-437c-aea5-cf0da50454a2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.976324] env[63293]: DEBUG nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.976502] env[63293]: DEBUG nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.976669] env[63293]: DEBUG nova.network.neutron [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.994604] env[63293]: DEBUG nova.network.neutron [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.019926] env[63293]: INFO nova.scheduler.client.report [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Deleted allocations for instance 1ac2f52a-27ef-4ce4-9540-288658f08344 [ 616.271369] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3a07f5-b5cc-42e6-bf27-890187d9c2fb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.279518] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "fd67940d-9552-42b3-b472-2ce0e7899ea6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.279518] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "fd67940d-9552-42b3-b472-2ce0e7899ea6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.283722] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dea6cb9-e509-4f60-9f35-706b304134fc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.316126] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6438a8d6-6492-4182-bc85-7ce50859c213 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.323483] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16f8927-5eb0-47ec-bd70-884452d8b396 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.338023] env[63293]: DEBUG nova.compute.provider_tree [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.500357] env[63293]: DEBUG nova.network.neutron [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.531784] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cca83848-d736-47e0-b2f7-b48eab712c6a tempest-ServersAdminTestJSON-1881773859 tempest-ServersAdminTestJSON-1881773859-project-member] Lock "1ac2f52a-27ef-4ce4-9540-288658f08344" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.128s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.843203] env[63293]: DEBUG nova.scheduler.client.report [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.005602] env[63293]: INFO nova.compute.manager [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] [instance: db725426-1b67-437c-aea5-cf0da50454a2] Took 1.03 seconds to deallocate network for instance. [ 617.033792] env[63293]: DEBUG nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 617.349474] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.350059] env[63293]: DEBUG nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 617.353618] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.908s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.562096] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.861143] env[63293]: DEBUG nova.compute.utils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.864753] env[63293]: DEBUG nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.865397] env[63293]: DEBUG nova.network.neutron [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 617.967434] env[63293]: DEBUG nova.policy [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '361d906209944261a239ffb1238d1514', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '088c4386156f43098c9401fd5975150e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 618.037541] env[63293]: INFO nova.scheduler.client.report [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Deleted allocations for instance db725426-1b67-437c-aea5-cf0da50454a2 [ 618.360476] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21eeb3fc-2196-4515-8689-cbc5f73efb74 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.366504] env[63293]: DEBUG nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 618.378195] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-079163d9-0804-435f-b0fc-bb5912392390 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.419798] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb0f604-a829-4d8c-a934-1c275a3b5637 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.427613] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc580a5-10d6-493d-9c97-e0e74a8362be {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.443421] env[63293]: DEBUG nova.compute.provider_tree [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.550732] env[63293]: DEBUG oslo_concurrency.lockutils [None req-16be822a-4ee4-4561-a8f9-639bb1882ae4 tempest-ListServerFiltersTestJSON-1153701687 tempest-ListServerFiltersTestJSON-1153701687-project-member] Lock "db725426-1b67-437c-aea5-cf0da50454a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.706s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.644519] env[63293]: DEBUG nova.network.neutron [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Successfully created port: 49767b23-0e55-492a-8ab0-6893c9580bda {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.946899] env[63293]: DEBUG nova.scheduler.client.report [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.057341] env[63293]: DEBUG nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 619.071107] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Acquiring lock "1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.071340] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Lock "1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.377902] env[63293]: DEBUG nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 619.404361] env[63293]: DEBUG nova.virt.hardware [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:33:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='319411438',id=19,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-272272225',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.405665] env[63293]: DEBUG nova.virt.hardware [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.405665] env[63293]: DEBUG nova.virt.hardware [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.405665] env[63293]: DEBUG nova.virt.hardware [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.405665] env[63293]: DEBUG nova.virt.hardware [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.405665] env[63293]: DEBUG nova.virt.hardware [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.405938] env[63293]: DEBUG nova.virt.hardware [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.405938] env[63293]: DEBUG nova.virt.hardware [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.405938] env[63293]: DEBUG nova.virt.hardware [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.405938] env[63293]: DEBUG nova.virt.hardware [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.406287] env[63293]: DEBUG nova.virt.hardware [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.407200] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099cc39c-5d01-4bb9-aa8b-908a9d209cc3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.416205] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e05db19-aedc-41f9-9e22-f5c9d90b91f9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.455653] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.103s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.456568] env[63293]: ERROR nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 53b6765d-312f-412f-9291-ac729b337036, please check neutron logs for more information. [ 619.456568] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Traceback (most recent call last): [ 619.456568] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.456568] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] self.driver.spawn(context, instance, image_meta, [ 619.456568] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 619.456568] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.456568] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.456568] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] vm_ref = self.build_virtual_machine(instance, [ 619.456568] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.456568] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.456568] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] for vif in network_info: [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] return self._sync_wrapper(fn, *args, **kwargs) [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] self.wait() [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] self[:] = self._gt.wait() [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] return self._exit_event.wait() [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] current.throw(*self._exc) [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.456899] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] result = function(*args, **kwargs) [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] return func(*args, **kwargs) [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] raise e [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] nwinfo = self.network_api.allocate_for_instance( [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] created_port_ids = self._update_ports_for_instance( [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] with excutils.save_and_reraise_exception(): [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] self.force_reraise() [ 619.459129] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.459476] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] raise self.value [ 619.459476] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.459476] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] updated_port = self._update_port( [ 619.459476] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.459476] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] _ensure_no_port_binding_failure(port) [ 619.459476] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.459476] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] raise exception.PortBindingFailed(port_id=port['id']) [ 619.459476] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] nova.exception.PortBindingFailed: Binding failed for port 53b6765d-312f-412f-9291-ac729b337036, please check neutron logs for more information. [ 619.459476] env[63293]: ERROR nova.compute.manager [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] [ 619.459476] env[63293]: DEBUG nova.compute.utils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Binding failed for port 53b6765d-312f-412f-9291-ac729b337036, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.459752] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.286s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.461929] env[63293]: DEBUG nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Build of instance 76ceec24-eebe-4b32-b77d-f57203881a94 was re-scheduled: Binding failed for port 53b6765d-312f-412f-9291-ac729b337036, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.462443] env[63293]: DEBUG nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.464200] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Acquiring lock "refresh_cache-76ceec24-eebe-4b32-b77d-f57203881a94" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.464200] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Acquired lock "refresh_cache-76ceec24-eebe-4b32-b77d-f57203881a94" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.464267] env[63293]: DEBUG nova.network.neutron [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.584233] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.792303] env[63293]: DEBUG nova.compute.manager [req-13985825-1646-416e-a966-76ae44575b2b req-7bb4b0bd-b645-4801-a3fc-cd2f02d0ce48 service nova] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Received event network-changed-49767b23-0e55-492a-8ab0-6893c9580bda {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 619.792554] env[63293]: DEBUG nova.compute.manager [req-13985825-1646-416e-a966-76ae44575b2b req-7bb4b0bd-b645-4801-a3fc-cd2f02d0ce48 service nova] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Refreshing instance network info cache due to event network-changed-49767b23-0e55-492a-8ab0-6893c9580bda. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 619.792699] env[63293]: DEBUG oslo_concurrency.lockutils [req-13985825-1646-416e-a966-76ae44575b2b req-7bb4b0bd-b645-4801-a3fc-cd2f02d0ce48 service nova] Acquiring lock "refresh_cache-a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.792838] env[63293]: DEBUG oslo_concurrency.lockutils [req-13985825-1646-416e-a966-76ae44575b2b req-7bb4b0bd-b645-4801-a3fc-cd2f02d0ce48 service nova] Acquired lock "refresh_cache-a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.792994] env[63293]: DEBUG nova.network.neutron [req-13985825-1646-416e-a966-76ae44575b2b req-7bb4b0bd-b645-4801-a3fc-cd2f02d0ce48 service nova] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Refreshing network info cache for port 49767b23-0e55-492a-8ab0-6893c9580bda {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 619.898671] env[63293]: ERROR nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 49767b23-0e55-492a-8ab0-6893c9580bda, please check neutron logs for more information. [ 619.898671] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 619.898671] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.898671] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 619.898671] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.898671] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 619.898671] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.898671] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 619.898671] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.898671] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 619.898671] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.898671] env[63293]: ERROR nova.compute.manager raise self.value [ 619.898671] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.898671] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 619.898671] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.898671] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 619.899091] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.899091] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 619.899091] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 49767b23-0e55-492a-8ab0-6893c9580bda, please check neutron logs for more information. [ 619.899091] env[63293]: ERROR nova.compute.manager [ 619.899091] env[63293]: Traceback (most recent call last): [ 619.899091] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 619.899091] env[63293]: listener.cb(fileno) [ 619.899091] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.899091] env[63293]: result = function(*args, **kwargs) [ 619.899091] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 619.899091] env[63293]: return func(*args, **kwargs) [ 619.899091] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.899091] env[63293]: raise e [ 619.899091] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.899091] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 619.899091] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.899091] env[63293]: created_port_ids = self._update_ports_for_instance( [ 619.899091] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.899091] env[63293]: with excutils.save_and_reraise_exception(): [ 619.899091] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.899091] env[63293]: self.force_reraise() [ 619.899091] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.899091] env[63293]: raise self.value [ 619.899091] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.899091] env[63293]: updated_port = self._update_port( [ 619.899091] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.899091] env[63293]: _ensure_no_port_binding_failure(port) [ 619.899091] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.899091] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 619.899825] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 49767b23-0e55-492a-8ab0-6893c9580bda, please check neutron logs for more information. [ 619.899825] env[63293]: Removing descriptor: 18 [ 619.899825] env[63293]: ERROR nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 49767b23-0e55-492a-8ab0-6893c9580bda, please check neutron logs for more information. [ 619.899825] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Traceback (most recent call last): [ 619.899825] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 619.899825] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] yield resources [ 619.899825] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.899825] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] self.driver.spawn(context, instance, image_meta, [ 619.899825] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 619.899825] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.899825] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.899825] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] vm_ref = self.build_virtual_machine(instance, [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] for vif in network_info: [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] return self._sync_wrapper(fn, *args, **kwargs) [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] self.wait() [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] self[:] = self._gt.wait() [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] return self._exit_event.wait() [ 619.900144] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] result = hub.switch() [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] return self.greenlet.switch() [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] result = function(*args, **kwargs) [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] return func(*args, **kwargs) [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] raise e [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] nwinfo = self.network_api.allocate_for_instance( [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.900467] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] created_port_ids = self._update_ports_for_instance( [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] with excutils.save_and_reraise_exception(): [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] self.force_reraise() [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] raise self.value [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] updated_port = self._update_port( [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] _ensure_no_port_binding_failure(port) [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.900824] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] raise exception.PortBindingFailed(port_id=port['id']) [ 619.901120] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] nova.exception.PortBindingFailed: Binding failed for port 49767b23-0e55-492a-8ab0-6893c9580bda, please check neutron logs for more information. [ 619.901120] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] [ 619.901120] env[63293]: INFO nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Terminating instance [ 619.903482] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Acquiring lock "refresh_cache-a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.986838] env[63293]: DEBUG nova.network.neutron [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.103055] env[63293]: DEBUG nova.network.neutron [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.313651] env[63293]: DEBUG nova.network.neutron [req-13985825-1646-416e-a966-76ae44575b2b req-7bb4b0bd-b645-4801-a3fc-cd2f02d0ce48 service nova] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.455605] env[63293]: DEBUG nova.network.neutron [req-13985825-1646-416e-a966-76ae44575b2b req-7bb4b0bd-b645-4801-a3fc-cd2f02d0ce48 service nova] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.461133] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded68a2d-34c5-4366-9c2e-a2e0d8075be5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.468259] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9884a4ad-4ee7-4f31-babb-4778e719d0b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.506113] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ae1d9d-8a1d-4e79-9f9e-7d397b2b7c3c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.513777] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b62f68-6689-418e-9c78-12c8bbf21901 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.529087] env[63293]: DEBUG nova.compute.provider_tree [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.608720] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Releasing lock "refresh_cache-76ceec24-eebe-4b32-b77d-f57203881a94" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.608953] env[63293]: DEBUG nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 620.609082] env[63293]: DEBUG nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.609301] env[63293]: DEBUG nova.network.neutron [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.631700] env[63293]: DEBUG nova.network.neutron [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.964688] env[63293]: DEBUG oslo_concurrency.lockutils [req-13985825-1646-416e-a966-76ae44575b2b req-7bb4b0bd-b645-4801-a3fc-cd2f02d0ce48 service nova] Releasing lock "refresh_cache-a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.965617] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Acquired lock "refresh_cache-a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.965617] env[63293]: DEBUG nova.network.neutron [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.035359] env[63293]: DEBUG nova.scheduler.client.report [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.135134] env[63293]: DEBUG nova.network.neutron [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.498700] env[63293]: DEBUG nova.network.neutron [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.541486] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.083s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.542207] env[63293]: ERROR nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c54cc217-14c7-43a7-9f9c-a980c39413ee, please check neutron logs for more information. [ 621.542207] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Traceback (most recent call last): [ 621.542207] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.542207] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] self.driver.spawn(context, instance, image_meta, [ 621.542207] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 621.542207] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.542207] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.542207] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] vm_ref = self.build_virtual_machine(instance, [ 621.542207] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.542207] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.542207] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] for vif in network_info: [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] return self._sync_wrapper(fn, *args, **kwargs) [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] self.wait() [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] self[:] = self._gt.wait() [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] return self._exit_event.wait() [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] result = hub.switch() [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.542525] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] return self.greenlet.switch() [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] result = function(*args, **kwargs) [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] return func(*args, **kwargs) [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] raise e [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] nwinfo = self.network_api.allocate_for_instance( [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] created_port_ids = self._update_ports_for_instance( [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] with excutils.save_and_reraise_exception(): [ 621.542869] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.543243] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] self.force_reraise() [ 621.543243] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.543243] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] raise self.value [ 621.543243] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.543243] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] updated_port = self._update_port( [ 621.543243] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.543243] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] _ensure_no_port_binding_failure(port) [ 621.543243] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.543243] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] raise exception.PortBindingFailed(port_id=port['id']) [ 621.543243] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] nova.exception.PortBindingFailed: Binding failed for port c54cc217-14c7-43a7-9f9c-a980c39413ee, please check neutron logs for more information. [ 621.543243] env[63293]: ERROR nova.compute.manager [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] [ 621.543538] env[63293]: DEBUG nova.compute.utils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Binding failed for port c54cc217-14c7-43a7-9f9c-a980c39413ee, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.548553] env[63293]: DEBUG nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Build of instance 25f0de66-914f-46b9-a15b-d936df5b3ab4 was re-scheduled: Binding failed for port c54cc217-14c7-43a7-9f9c-a980c39413ee, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 621.548988] env[63293]: DEBUG nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 621.549328] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Acquiring lock "refresh_cache-25f0de66-914f-46b9-a15b-d936df5b3ab4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.549387] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Acquired lock "refresh_cache-25f0de66-914f-46b9-a15b-d936df5b3ab4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.549538] env[63293]: DEBUG nova.network.neutron [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.550585] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.597s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.639561] env[63293]: INFO nova.compute.manager [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] [instance: 76ceec24-eebe-4b32-b77d-f57203881a94] Took 1.03 seconds to deallocate network for instance. [ 621.773815] env[63293]: DEBUG nova.network.neutron [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.888079] env[63293]: DEBUG nova.compute.manager [req-5e64c85f-445c-4832-9d51-2ff190c1a847 req-6e2a47f0-c648-492e-95d0-af34e4ed781e service nova] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Received event network-vif-deleted-49767b23-0e55-492a-8ab0-6893c9580bda {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.085484] env[63293]: DEBUG nova.network.neutron [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.232583] env[63293]: DEBUG nova.network.neutron [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.276295] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Releasing lock "refresh_cache-a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.278444] env[63293]: DEBUG nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 622.278444] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.278444] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb2d6f91-e143-4e7a-872b-dd02f3cb3049 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.286393] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa060593-35d7-487c-9652-edd15fb728df {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.309926] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a492b0a0-0bf1-4e64-bd51-ad131cd17a3b could not be found. [ 622.310079] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 622.310261] env[63293]: INFO nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 622.310505] env[63293]: DEBUG oslo.service.loopingcall [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.313140] env[63293]: DEBUG nova.compute.manager [-] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.313275] env[63293]: DEBUG nova.network.neutron [-] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.350929] env[63293]: DEBUG nova.network.neutron [-] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.548050] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96bbfc7-f2a7-45eb-b9b9-cf5424bd739e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.178757] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Releasing lock "refresh_cache-25f0de66-914f-46b9-a15b-d936df5b3ab4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.178919] env[63293]: DEBUG nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 623.181201] env[63293]: DEBUG nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.181201] env[63293]: DEBUG nova.network.neutron [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.181201] env[63293]: DEBUG nova.network.neutron [-] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.183164] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9609da8a-c8b3-40e9-bf74-0c0cb00bc61b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.215211] env[63293]: INFO nova.scheduler.client.report [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Deleted allocations for instance 76ceec24-eebe-4b32-b77d-f57203881a94 [ 623.221930] env[63293]: DEBUG nova.network.neutron [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.223113] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d33e6cf6-83bd-4626-b1e2-b507f932a5d4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.232950] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e89ce3-5d76-4491-a3f9-b211b1d505c6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.246676] env[63293]: DEBUG nova.compute.provider_tree [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.684242] env[63293]: INFO nova.compute.manager [-] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Took 1.37 seconds to deallocate network for instance. [ 623.686527] env[63293]: DEBUG nova.compute.claims [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 623.686624] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.730710] env[63293]: DEBUG nova.network.neutron [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.730710] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9745712e-eb69-4a97-8613-bdcae0991087 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453 tempest-FloatingIPsAssociationNegativeTestJSON-1146749453-project-member] Lock "76ceec24-eebe-4b32-b77d-f57203881a94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.742s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.751753] env[63293]: DEBUG nova.scheduler.client.report [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.233619] env[63293]: INFO nova.compute.manager [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] [instance: 25f0de66-914f-46b9-a15b-d936df5b3ab4] Took 1.05 seconds to deallocate network for instance. [ 624.238298] env[63293]: DEBUG nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.259761] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.707s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.259761] env[63293]: ERROR nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ddc895f1-d2a2-4a78-8e01-0c6b6660dee7, please check neutron logs for more information. [ 624.259761] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Traceback (most recent call last): [ 624.259761] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.259761] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] self.driver.spawn(context, instance, image_meta, [ 624.259761] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 624.259761] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.259761] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.259761] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] vm_ref = self.build_virtual_machine(instance, [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] for vif in network_info: [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] return self._sync_wrapper(fn, *args, **kwargs) [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] self.wait() [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] self[:] = self._gt.wait() [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] return self._exit_event.wait() [ 624.260021] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] result = hub.switch() [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] return self.greenlet.switch() [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] result = function(*args, **kwargs) [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] return func(*args, **kwargs) [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] raise e [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] nwinfo = self.network_api.allocate_for_instance( [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.260340] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] created_port_ids = self._update_ports_for_instance( [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] with excutils.save_and_reraise_exception(): [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] self.force_reraise() [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] raise self.value [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] updated_port = self._update_port( [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] _ensure_no_port_binding_failure(port) [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.260671] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] raise exception.PortBindingFailed(port_id=port['id']) [ 624.260961] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] nova.exception.PortBindingFailed: Binding failed for port ddc895f1-d2a2-4a78-8e01-0c6b6660dee7, please check neutron logs for more information. [ 624.260961] env[63293]: ERROR nova.compute.manager [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] [ 624.260961] env[63293]: DEBUG nova.compute.utils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Binding failed for port ddc895f1-d2a2-4a78-8e01-0c6b6660dee7, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 624.262024] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.922s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.265071] env[63293]: DEBUG nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Build of instance c1f9f075-7d1c-4c05-9dfd-43820d2d07ad was re-scheduled: Binding failed for port ddc895f1-d2a2-4a78-8e01-0c6b6660dee7, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 624.265342] env[63293]: DEBUG nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 624.265956] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Acquiring lock "refresh_cache-c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.265956] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Acquired lock "refresh_cache-c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.265956] env[63293]: DEBUG nova.network.neutron [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.776951] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.850135] env[63293]: DEBUG nova.network.neutron [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.021830] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Acquiring lock "985a5c30-551f-4de7-bc85-d53f002e0f72" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.022184] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Lock "985a5c30-551f-4de7-bc85-d53f002e0f72" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.044675] env[63293]: DEBUG nova.network.neutron [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.232880] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db58df35-c827-41de-9b2d-3d5e70903e52 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.241612] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3ed1b1-cbbf-48ca-aa13-4c178aabf3f0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.279208] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7042a7b5-8038-46b9-8fa1-6aa85bb00816 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.287281] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349aed91-72c0-4dc3-b821-fe6cca65e2dc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.305688] env[63293]: DEBUG nova.compute.provider_tree [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.312020] env[63293]: INFO nova.scheduler.client.report [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Deleted allocations for instance 25f0de66-914f-46b9-a15b-d936df5b3ab4 [ 625.547949] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Releasing lock "refresh_cache-c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.549009] env[63293]: DEBUG nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 625.549009] env[63293]: DEBUG nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.549009] env[63293]: DEBUG nova.network.neutron [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.581252] env[63293]: DEBUG nova.network.neutron [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.814928] env[63293]: DEBUG nova.scheduler.client.report [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.818616] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c1d08ca2-3e02-463e-b8bf-253ba53ba33f tempest-ImagesOneServerNegativeTestJSON-661176166 tempest-ImagesOneServerNegativeTestJSON-661176166-project-member] Lock "25f0de66-914f-46b9-a15b-d936df5b3ab4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.597s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.088028] env[63293]: DEBUG nova.network.neutron [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.324552] env[63293]: DEBUG nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 626.334940] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.073s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.339015] env[63293]: ERROR nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bfb28db2-e515-4222-9f85-aace51153a77, please check neutron logs for more information. [ 626.339015] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Traceback (most recent call last): [ 626.339015] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.339015] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] self.driver.spawn(context, instance, image_meta, [ 626.339015] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 626.339015] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.339015] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.339015] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] vm_ref = self.build_virtual_machine(instance, [ 626.339015] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.339015] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.339015] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] for vif in network_info: [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] return self._sync_wrapper(fn, *args, **kwargs) [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] self.wait() [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] self[:] = self._gt.wait() [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] return self._exit_event.wait() [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] result = hub.switch() [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.339407] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] return self.greenlet.switch() [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] result = function(*args, **kwargs) [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] return func(*args, **kwargs) [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] raise e [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] nwinfo = self.network_api.allocate_for_instance( [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] created_port_ids = self._update_ports_for_instance( [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] with excutils.save_and_reraise_exception(): [ 626.339719] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.340081] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] self.force_reraise() [ 626.340081] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.340081] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] raise self.value [ 626.340081] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.340081] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] updated_port = self._update_port( [ 626.340081] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.340081] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] _ensure_no_port_binding_failure(port) [ 626.340081] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.340081] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] raise exception.PortBindingFailed(port_id=port['id']) [ 626.340081] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] nova.exception.PortBindingFailed: Binding failed for port bfb28db2-e515-4222-9f85-aace51153a77, please check neutron logs for more information. [ 626.340081] env[63293]: ERROR nova.compute.manager [instance: a236be17-8965-4785-834e-9ce22eeb75eb] [ 626.340331] env[63293]: DEBUG nova.compute.utils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Binding failed for port bfb28db2-e515-4222-9f85-aace51153a77, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 626.340331] env[63293]: DEBUG nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Build of instance a236be17-8965-4785-834e-9ce22eeb75eb was re-scheduled: Binding failed for port bfb28db2-e515-4222-9f85-aace51153a77, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 626.340759] env[63293]: DEBUG nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 626.340759] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Acquiring lock "refresh_cache-a236be17-8965-4785-834e-9ce22eeb75eb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.340891] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Acquired lock "refresh_cache-a236be17-8965-4785-834e-9ce22eeb75eb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.341025] env[63293]: DEBUG nova.network.neutron [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.345621] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.614s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.347826] env[63293]: INFO nova.compute.claims [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.594072] env[63293]: INFO nova.compute.manager [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] [instance: c1f9f075-7d1c-4c05-9dfd-43820d2d07ad] Took 1.04 seconds to deallocate network for instance. [ 626.879338] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.882028] env[63293]: DEBUG nova.network.neutron [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.005901] env[63293]: DEBUG nova.network.neutron [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.510332] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Releasing lock "refresh_cache-a236be17-8965-4785-834e-9ce22eeb75eb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.510332] env[63293]: DEBUG nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 627.510332] env[63293]: DEBUG nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.512912] env[63293]: DEBUG nova.network.neutron [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 627.535857] env[63293]: DEBUG nova.network.neutron [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.648537] env[63293]: INFO nova.scheduler.client.report [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Deleted allocations for instance c1f9f075-7d1c-4c05-9dfd-43820d2d07ad [ 627.909472] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b58850-2cdd-4742-8178-e18a394247b9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.917042] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524b55af-b0ac-42a7-a381-69d2fc8f7fe4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.955825] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02e1199-eeb7-45d6-9394-37b524585b60 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.967017] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29854b6f-d72d-4f06-93da-8712b2bf399a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.979573] env[63293]: DEBUG nova.compute.provider_tree [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.041939] env[63293]: DEBUG nova.network.neutron [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.161515] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b60d2898-2342-4459-83e9-efe07af01910 tempest-VolumesAssistedSnapshotsTest-2141301124 tempest-VolumesAssistedSnapshotsTest-2141301124-project-member] Lock "c1f9f075-7d1c-4c05-9dfd-43820d2d07ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.142s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.482932] env[63293]: DEBUG nova.scheduler.client.report [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.544481] env[63293]: INFO nova.compute.manager [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] [instance: a236be17-8965-4785-834e-9ce22eeb75eb] Took 1.03 seconds to deallocate network for instance. [ 628.666360] env[63293]: DEBUG nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 628.995036] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.647s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.995036] env[63293]: DEBUG nova.compute.manager [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 628.996336] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.331s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.000514] env[63293]: INFO nova.compute.claims [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.195922] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.508794] env[63293]: DEBUG nova.compute.utils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 629.511459] env[63293]: DEBUG nova.compute.manager [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Not allocating networking since 'none' was specified. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 629.590412] env[63293]: INFO nova.scheduler.client.report [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Deleted allocations for instance a236be17-8965-4785-834e-9ce22eeb75eb [ 630.016026] env[63293]: DEBUG nova.compute.manager [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 630.099976] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8ea6e7e9-7778-4064-ac67-d14432df04c8 tempest-AttachInterfacesV270Test-485999998 tempest-AttachInterfacesV270Test-485999998-project-member] Lock "a236be17-8965-4785-834e-9ce22eeb75eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.950s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.560582] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7355eb-cdfa-4584-b455-a7a3ccd383a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.572247] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffe5ca2-50cd-4f02-bdb3-9c0aed9a8ab7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.608372] env[63293]: DEBUG nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 630.611626] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183f2991-057d-4ef1-bf79-71e8cdeae47a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.623031] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1a8819-7cfb-4baf-9295-f8b6f18209b7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.638067] env[63293]: DEBUG nova.compute.provider_tree [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.033018] env[63293]: DEBUG nova.compute.manager [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 631.063310] env[63293]: DEBUG nova.virt.hardware [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 631.063310] env[63293]: DEBUG nova.virt.hardware [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 631.063310] env[63293]: DEBUG nova.virt.hardware [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.063499] env[63293]: DEBUG nova.virt.hardware [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 631.063499] env[63293]: DEBUG nova.virt.hardware [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.064209] env[63293]: DEBUG nova.virt.hardware [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 631.064652] env[63293]: DEBUG nova.virt.hardware [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 631.064986] env[63293]: DEBUG nova.virt.hardware [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 631.065275] env[63293]: DEBUG nova.virt.hardware [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 631.065572] env[63293]: DEBUG nova.virt.hardware [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 631.065848] env[63293]: DEBUG nova.virt.hardware [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.067234] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ba764a-a893-402f-bd3d-58b36b15bca9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.077366] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c75a77-79cb-45da-a446-06ea9d7ddae8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.094293] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 631.100289] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Creating folder: Project (25f5b74a09a54a8098b167225d1d3d04). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 631.100621] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-713f19a7-4d22-4fcf-80ab-649e37a7833b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.110647] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Created folder: Project (25f5b74a09a54a8098b167225d1d3d04) in parent group-v283678. [ 631.110851] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Creating folder: Instances. Parent ref: group-v283687. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 631.113573] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6dfa126c-8685-4582-a7d5-31d6e40eeff9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.126358] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Created folder: Instances in parent group-v283687. [ 631.126645] env[63293]: DEBUG oslo.service.loopingcall [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.126861] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 631.127078] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71de27e2-fe99-4b32-bd91-c9ad75c58b25 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.145774] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.145774] env[63293]: DEBUG nova.scheduler.client.report [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.153596] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 631.153596] env[63293]: value = "task-1327448" [ 631.153596] env[63293]: _type = "Task" [ 631.153596] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.161763] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327448, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.419422] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 631.419689] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 631.650028] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.653s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.650742] env[63293]: DEBUG nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 631.655116] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.016s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.658173] env[63293]: INFO nova.compute.claims [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.674648] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327448, 'name': CreateVM_Task, 'duration_secs': 0.272085} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.674984] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 631.675467] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.675665] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.676033] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 631.678368] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50c576c4-ac09-41d5-af99-4206ae667aaf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.684169] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for the task: (returnval){ [ 631.684169] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a2ffe7-da53-9c0b-7cf0-9a3a70224176" [ 631.684169] env[63293]: _type = "Task" [ 631.684169] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.690112] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a2ffe7-da53-9c0b-7cf0-9a3a70224176, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.929519] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 631.929519] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Starting heal instance info cache {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 631.929519] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Rebuilding the list of instances to heal {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 632.169495] env[63293]: DEBUG nova.compute.utils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 632.169495] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquiring lock "b5334832-25dc-4b03-9689-388a1382200a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.169495] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "b5334832-25dc-4b03-9689-388a1382200a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.169495] env[63293]: DEBUG nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 632.169898] env[63293]: DEBUG nova.network.neutron [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 632.201299] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a2ffe7-da53-9c0b-7cf0-9a3a70224176, 'name': SearchDatastore_Task, 'duration_secs': 0.011242} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.203039] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.203130] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 632.203395] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.203511] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.203693] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 632.203959] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dff09daf-d434-4944-adae-0d20baa55aca {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.214188] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 632.214372] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 632.215399] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-036789d9-54e1-4d72-9475-f276ea543bf5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.221496] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for the task: (returnval){ [ 632.221496] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523025b9-666c-5575-de20-a893189f8a49" [ 632.221496] env[63293]: _type = "Task" [ 632.221496] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.229839] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523025b9-666c-5575-de20-a893189f8a49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.291796] env[63293]: DEBUG nova.policy [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b71c6c0cc133419ea1df4ff2ebba1d8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b67ae2932d741c4b3a03d5d344127ef', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 632.436673] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 632.436673] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 632.436812] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 632.436926] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 632.437066] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Didn't find any instances for network info cache update. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 632.438240] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 632.438728] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 632.438728] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 632.438866] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 632.439253] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 632.439253] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 632.439327] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63293) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 632.439423] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 632.678970] env[63293]: DEBUG nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 632.738685] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523025b9-666c-5575-de20-a893189f8a49, 'name': SearchDatastore_Task, 'duration_secs': 0.008798} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.739658] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd0d757e-8846-44ec-b962-6f30b40df911 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.744736] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for the task: (returnval){ [ 632.744736] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523e83ed-c556-7d10-f2aa-0e66ace6f978" [ 632.744736] env[63293]: _type = "Task" [ 632.744736] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.752458] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523e83ed-c556-7d10-f2aa-0e66ace6f978, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.946472] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.098304] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5898217-2deb-4c30-9a05-002dd5ae867a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.114300] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75705fdb-fa5b-4098-85ca-3b685e8e0f33 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.153926] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e16c30-280f-4278-b335-71fcd3481b44 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.165192] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9461f77f-f0a3-41b1-b6a0-5fcfd342b86a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.183030] env[63293]: DEBUG nova.compute.provider_tree [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.201232] env[63293]: DEBUG nova.network.neutron [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Successfully created port: 4993189a-a876-4140-8985-d263769da28b {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.256676] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523e83ed-c556-7d10-f2aa-0e66ace6f978, 'name': SearchDatastore_Task, 'duration_secs': 0.009219} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.257024] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.257314] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] b2055187-edd8-4a8c-9811-183cf7ba3b39/b2055187-edd8-4a8c-9811-183cf7ba3b39.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 633.257541] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-473046c1-133f-406f-8b93-7419a6dda226 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.264782] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for the task: (returnval){ [ 633.264782] env[63293]: value = "task-1327449" [ 633.264782] env[63293]: _type = "Task" [ 633.264782] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.275569] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327449, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.687414] env[63293]: DEBUG nova.scheduler.client.report [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.695726] env[63293]: DEBUG nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 633.732631] env[63293]: DEBUG nova.virt.hardware [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 633.732965] env[63293]: DEBUG nova.virt.hardware [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 633.733163] env[63293]: DEBUG nova.virt.hardware [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.733383] env[63293]: DEBUG nova.virt.hardware [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 633.733569] env[63293]: DEBUG nova.virt.hardware [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.733670] env[63293]: DEBUG nova.virt.hardware [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 633.733880] env[63293]: DEBUG nova.virt.hardware [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 633.734050] env[63293]: DEBUG nova.virt.hardware [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 633.734222] env[63293]: DEBUG nova.virt.hardware [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 633.734384] env[63293]: DEBUG nova.virt.hardware [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 633.734557] env[63293]: DEBUG nova.virt.hardware [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.735978] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178cd83b-e25d-4d90-88dd-0f5095b05627 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.746343] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8248ab-305f-4a10-9ff9-10b8d7effba3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.775484] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327449, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.200020] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.200020] env[63293]: DEBUG nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 634.201212] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.639s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.203078] env[63293]: INFO nova.compute.claims [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 634.278830] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327449, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.560764] env[63293]: DEBUG nova.compute.manager [req-60cbf80f-beab-4b4f-9e27-868fa66a86c6 req-9157427b-2133-4059-b66d-01617b03d6ff service nova] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Received event network-changed-4993189a-a876-4140-8985-d263769da28b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.560956] env[63293]: DEBUG nova.compute.manager [req-60cbf80f-beab-4b4f-9e27-868fa66a86c6 req-9157427b-2133-4059-b66d-01617b03d6ff service nova] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Refreshing instance network info cache due to event network-changed-4993189a-a876-4140-8985-d263769da28b. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 634.561440] env[63293]: DEBUG oslo_concurrency.lockutils [req-60cbf80f-beab-4b4f-9e27-868fa66a86c6 req-9157427b-2133-4059-b66d-01617b03d6ff service nova] Acquiring lock "refresh_cache-510b9ea6-04f1-4704-b5e4-e19db57769a8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.561440] env[63293]: DEBUG oslo_concurrency.lockutils [req-60cbf80f-beab-4b4f-9e27-868fa66a86c6 req-9157427b-2133-4059-b66d-01617b03d6ff service nova] Acquired lock "refresh_cache-510b9ea6-04f1-4704-b5e4-e19db57769a8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.561550] env[63293]: DEBUG nova.network.neutron [req-60cbf80f-beab-4b4f-9e27-868fa66a86c6 req-9157427b-2133-4059-b66d-01617b03d6ff service nova] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Refreshing network info cache for port 4993189a-a876-4140-8985-d263769da28b {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 634.709560] env[63293]: DEBUG nova.compute.utils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.713011] env[63293]: DEBUG nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.713011] env[63293]: DEBUG nova.network.neutron [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 634.780262] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327449, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.801744] env[63293]: DEBUG nova.policy [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af33165b77294edfbffc57caec100d3f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a3bcac63f7844e10a166eca717f910a1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.923782] env[63293]: ERROR nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4993189a-a876-4140-8985-d263769da28b, please check neutron logs for more information. [ 634.923782] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 634.923782] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.923782] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 634.923782] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.923782] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 634.923782] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.923782] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 634.923782] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.923782] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 634.923782] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.923782] env[63293]: ERROR nova.compute.manager raise self.value [ 634.923782] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.923782] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 634.923782] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.923782] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 634.924419] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.924419] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 634.924419] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4993189a-a876-4140-8985-d263769da28b, please check neutron logs for more information. [ 634.924419] env[63293]: ERROR nova.compute.manager [ 634.924419] env[63293]: Traceback (most recent call last): [ 634.924419] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 634.924419] env[63293]: listener.cb(fileno) [ 634.924419] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.924419] env[63293]: result = function(*args, **kwargs) [ 634.924419] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 634.924419] env[63293]: return func(*args, **kwargs) [ 634.924419] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.924419] env[63293]: raise e [ 634.924419] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.924419] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 634.924419] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.924419] env[63293]: created_port_ids = self._update_ports_for_instance( [ 634.924419] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.924419] env[63293]: with excutils.save_and_reraise_exception(): [ 634.924419] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.924419] env[63293]: self.force_reraise() [ 634.924419] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.924419] env[63293]: raise self.value [ 634.924419] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.924419] env[63293]: updated_port = self._update_port( [ 634.924419] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.924419] env[63293]: _ensure_no_port_binding_failure(port) [ 634.924419] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.924419] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.925624] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 4993189a-a876-4140-8985-d263769da28b, please check neutron logs for more information. [ 634.925624] env[63293]: Removing descriptor: 18 [ 634.925624] env[63293]: ERROR nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4993189a-a876-4140-8985-d263769da28b, please check neutron logs for more information. [ 634.925624] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Traceback (most recent call last): [ 634.925624] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 634.925624] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] yield resources [ 634.925624] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.925624] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] self.driver.spawn(context, instance, image_meta, [ 634.925624] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 634.925624] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.925624] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.925624] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] vm_ref = self.build_virtual_machine(instance, [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] for vif in network_info: [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] return self._sync_wrapper(fn, *args, **kwargs) [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] self.wait() [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] self[:] = self._gt.wait() [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] return self._exit_event.wait() [ 634.926144] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] result = hub.switch() [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] return self.greenlet.switch() [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] result = function(*args, **kwargs) [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] return func(*args, **kwargs) [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] raise e [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] nwinfo = self.network_api.allocate_for_instance( [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.926550] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] created_port_ids = self._update_ports_for_instance( [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] with excutils.save_and_reraise_exception(): [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] self.force_reraise() [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] raise self.value [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] updated_port = self._update_port( [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] _ensure_no_port_binding_failure(port) [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.926985] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] raise exception.PortBindingFailed(port_id=port['id']) [ 634.927942] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] nova.exception.PortBindingFailed: Binding failed for port 4993189a-a876-4140-8985-d263769da28b, please check neutron logs for more information. [ 634.927942] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] [ 634.927942] env[63293]: INFO nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Terminating instance [ 634.928089] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "refresh_cache-510b9ea6-04f1-4704-b5e4-e19db57769a8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.107443] env[63293]: DEBUG nova.network.neutron [req-60cbf80f-beab-4b4f-9e27-868fa66a86c6 req-9157427b-2133-4059-b66d-01617b03d6ff service nova] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.213463] env[63293]: DEBUG nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.279602] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327449, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.548059} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.283023] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] b2055187-edd8-4a8c-9811-183cf7ba3b39/b2055187-edd8-4a8c-9811-183cf7ba3b39.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 635.283454] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 635.284039] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-02a7267d-4be3-4657-a704-4874622bb470 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.294907] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for the task: (returnval){ [ 635.294907] env[63293]: value = "task-1327451" [ 635.294907] env[63293]: _type = "Task" [ 635.294907] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.306650] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327451, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.536676] env[63293]: DEBUG nova.network.neutron [req-60cbf80f-beab-4b4f-9e27-868fa66a86c6 req-9157427b-2133-4059-b66d-01617b03d6ff service nova] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.605142] env[63293]: DEBUG nova.network.neutron [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Successfully created port: 453608e9-ae55-48dd-b58b-6189194d69c0 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 635.671129] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4f8ec6-508a-4cb2-befc-ffcf9e66ab89 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.679713] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7542a2-813e-4d57-a560-3f0ea34dafa5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.710235] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bcf35cf-302e-4f5a-b014-6b264a741351 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.718500] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1453a7cb-a33d-47bf-b50f-5cf1036307ff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.744997] env[63293]: DEBUG nova.compute.provider_tree [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.809441] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327451, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065799} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.809596] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 635.810668] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba361e2-14a7-4ef9-885b-d067f5f208c6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.836626] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Reconfiguring VM instance instance-00000016 to attach disk [datastore1] b2055187-edd8-4a8c-9811-183cf7ba3b39/b2055187-edd8-4a8c-9811-183cf7ba3b39.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 635.836953] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2694121c-6f13-4547-99f5-fdf14ac22828 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.859289] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for the task: (returnval){ [ 635.859289] env[63293]: value = "task-1327452" [ 635.859289] env[63293]: _type = "Task" [ 635.859289] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.868030] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327452, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.045178] env[63293]: DEBUG oslo_concurrency.lockutils [req-60cbf80f-beab-4b4f-9e27-868fa66a86c6 req-9157427b-2133-4059-b66d-01617b03d6ff service nova] Releasing lock "refresh_cache-510b9ea6-04f1-4704-b5e4-e19db57769a8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.045854] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquired lock "refresh_cache-510b9ea6-04f1-4704-b5e4-e19db57769a8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.046083] env[63293]: DEBUG nova.network.neutron [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.230710] env[63293]: DEBUG nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 636.250688] env[63293]: DEBUG nova.scheduler.client.report [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.274146] env[63293]: DEBUG nova.virt.hardware [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.274416] env[63293]: DEBUG nova.virt.hardware [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.275105] env[63293]: DEBUG nova.virt.hardware [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.275344] env[63293]: DEBUG nova.virt.hardware [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.275579] env[63293]: DEBUG nova.virt.hardware [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.275805] env[63293]: DEBUG nova.virt.hardware [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.276043] env[63293]: DEBUG nova.virt.hardware [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.276209] env[63293]: DEBUG nova.virt.hardware [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.276382] env[63293]: DEBUG nova.virt.hardware [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.276766] env[63293]: DEBUG nova.virt.hardware [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.276766] env[63293]: DEBUG nova.virt.hardware [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.277613] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45ec0f5-1696-4728-9288-ca47227c1135 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.290079] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09cc01fb-63b3-4af3-bf1c-b1b8839cdfa0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.370845] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327452, 'name': ReconfigVM_Task, 'duration_secs': 0.301962} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.371335] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Reconfigured VM instance instance-00000016 to attach disk [datastore1] b2055187-edd8-4a8c-9811-183cf7ba3b39/b2055187-edd8-4a8c-9811-183cf7ba3b39.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 636.372214] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b63a2ff2-7663-4300-9862-07a001464963 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.383899] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for the task: (returnval){ [ 636.383899] env[63293]: value = "task-1327454" [ 636.383899] env[63293]: _type = "Task" [ 636.383899] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.400133] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327454, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.595018] env[63293]: DEBUG nova.compute.manager [req-f62ca2da-ccce-4369-8517-3b83394b17d3 req-55812e6e-9b53-4082-abb1-721660ae2ea4 service nova] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Received event network-vif-deleted-4993189a-a876-4140-8985-d263769da28b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 636.595733] env[63293]: DEBUG nova.network.neutron [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.761022] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.761022] env[63293]: DEBUG nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 636.762842] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.179s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.765173] env[63293]: INFO nova.compute.claims [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 636.775066] env[63293]: DEBUG nova.network.neutron [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.898201] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327454, 'name': Rename_Task, 'duration_secs': 0.193522} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.901018] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 636.901018] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9edb8be2-076a-41e5-9cab-759973882a1e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.909017] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for the task: (returnval){ [ 636.909017] env[63293]: value = "task-1327455" [ 636.909017] env[63293]: _type = "Task" [ 636.909017] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.917951] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327455, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.272535] env[63293]: DEBUG nova.compute.utils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 637.276051] env[63293]: DEBUG nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 637.276237] env[63293]: DEBUG nova.network.neutron [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 637.278489] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Releasing lock "refresh_cache-510b9ea6-04f1-4704-b5e4-e19db57769a8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.278865] env[63293]: DEBUG nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 637.279074] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 637.279349] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1accb8b0-05e3-42e1-b9de-bca9d19dfe5d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.290749] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe447dd-f133-4e8d-918f-baa14b6f0779 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.318631] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 510b9ea6-04f1-4704-b5e4-e19db57769a8 could not be found. [ 637.318631] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.318771] env[63293]: INFO nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 637.319041] env[63293]: DEBUG oslo.service.loopingcall [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.320705] env[63293]: DEBUG nova.compute.manager [-] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.320705] env[63293]: DEBUG nova.network.neutron [-] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.335431] env[63293]: DEBUG nova.network.neutron [-] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.422270] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327455, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.440204] env[63293]: DEBUG nova.policy [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc557354560049119144c62dd61d0816', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2af826d819b4cc4ab27ff8b95144039', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 637.777469] env[63293]: DEBUG nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 637.840288] env[63293]: DEBUG nova.network.neutron [-] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.930088] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327455, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.281159] env[63293]: ERROR nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 453608e9-ae55-48dd-b58b-6189194d69c0, please check neutron logs for more information. [ 638.281159] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.281159] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.281159] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.281159] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.281159] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.281159] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.281159] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.281159] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.281159] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 638.281159] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.281159] env[63293]: ERROR nova.compute.manager raise self.value [ 638.281159] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.281159] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.281159] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.281159] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.281840] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.281840] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.281840] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 453608e9-ae55-48dd-b58b-6189194d69c0, please check neutron logs for more information. [ 638.281840] env[63293]: ERROR nova.compute.manager [ 638.285379] env[63293]: Traceback (most recent call last): [ 638.285379] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.285379] env[63293]: listener.cb(fileno) [ 638.285379] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.285379] env[63293]: result = function(*args, **kwargs) [ 638.285379] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.285379] env[63293]: return func(*args, **kwargs) [ 638.285379] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.285379] env[63293]: raise e [ 638.285379] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.285379] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 638.285379] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.285379] env[63293]: created_port_ids = self._update_ports_for_instance( [ 638.285379] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.285379] env[63293]: with excutils.save_and_reraise_exception(): [ 638.285379] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.285379] env[63293]: self.force_reraise() [ 638.285379] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.285379] env[63293]: raise self.value [ 638.285379] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.285379] env[63293]: updated_port = self._update_port( [ 638.285379] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.285379] env[63293]: _ensure_no_port_binding_failure(port) [ 638.285379] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.285379] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.285379] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 453608e9-ae55-48dd-b58b-6189194d69c0, please check neutron logs for more information. [ 638.285379] env[63293]: Removing descriptor: 17 [ 638.286128] env[63293]: ERROR nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 453608e9-ae55-48dd-b58b-6189194d69c0, please check neutron logs for more information. [ 638.286128] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Traceback (most recent call last): [ 638.286128] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 638.286128] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] yield resources [ 638.286128] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.286128] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] self.driver.spawn(context, instance, image_meta, [ 638.286128] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 638.286128] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.286128] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.286128] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] vm_ref = self.build_virtual_machine(instance, [ 638.286128] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] for vif in network_info: [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] return self._sync_wrapper(fn, *args, **kwargs) [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] self.wait() [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] self[:] = self._gt.wait() [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] return self._exit_event.wait() [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.286580] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] result = hub.switch() [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] return self.greenlet.switch() [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] result = function(*args, **kwargs) [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] return func(*args, **kwargs) [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] raise e [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] nwinfo = self.network_api.allocate_for_instance( [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] created_port_ids = self._update_ports_for_instance( [ 638.286995] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] with excutils.save_and_reraise_exception(): [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] self.force_reraise() [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] raise self.value [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] updated_port = self._update_port( [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] _ensure_no_port_binding_failure(port) [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] raise exception.PortBindingFailed(port_id=port['id']) [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] nova.exception.PortBindingFailed: Binding failed for port 453608e9-ae55-48dd-b58b-6189194d69c0, please check neutron logs for more information. [ 638.287628] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] [ 638.288117] env[63293]: INFO nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Terminating instance [ 638.294038] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "refresh_cache-eaaf2515-cc4b-42c7-aacd-14074c760924" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.294038] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquired lock "refresh_cache-eaaf2515-cc4b-42c7-aacd-14074c760924" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.294038] env[63293]: DEBUG nova.network.neutron [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.332964] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437e9b85-ba8e-4266-86fe-7a5937969b59 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.343378] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3460f3-cdb5-4bdf-9983-ee609222fd3e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.350881] env[63293]: INFO nova.compute.manager [-] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Took 1.03 seconds to deallocate network for instance. [ 638.353880] env[63293]: DEBUG nova.compute.claims [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.354105] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.387821] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c8528db-af82-4341-853d-2daff9cf75ac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.399277] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45888f2d-548b-4051-85b6-898e94d45704 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.416903] env[63293]: DEBUG nova.compute.provider_tree [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.430164] env[63293]: DEBUG oslo_vmware.api [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327455, 'name': PowerOnVM_Task, 'duration_secs': 1.048499} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.430252] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 638.430462] env[63293]: INFO nova.compute.manager [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Took 7.40 seconds to spawn the instance on the hypervisor. [ 638.430639] env[63293]: DEBUG nova.compute.manager [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 638.434701] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5402cc0e-d0cc-45ec-85ed-be148450e82f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.569376] env[63293]: DEBUG nova.network.neutron [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Successfully created port: 255fd7f1-f9a0-4469-80e0-980345a3cb6d {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 638.799468] env[63293]: DEBUG nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 638.839635] env[63293]: DEBUG nova.network.neutron [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.848573] env[63293]: DEBUG nova.virt.hardware [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 638.850260] env[63293]: DEBUG nova.virt.hardware [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 638.850260] env[63293]: DEBUG nova.virt.hardware [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.850260] env[63293]: DEBUG nova.virt.hardware [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 638.850260] env[63293]: DEBUG nova.virt.hardware [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.850260] env[63293]: DEBUG nova.virt.hardware [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 638.850260] env[63293]: DEBUG nova.virt.hardware [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 638.850545] env[63293]: DEBUG nova.virt.hardware [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 638.850545] env[63293]: DEBUG nova.virt.hardware [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 638.850545] env[63293]: DEBUG nova.virt.hardware [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 638.851257] env[63293]: DEBUG nova.virt.hardware [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 638.852716] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc391d8-d27f-4e7c-9497-a94d30cd5ccb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.871019] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d8ca74d-be4f-479f-b991-5bc289d233b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.923476] env[63293]: DEBUG nova.scheduler.client.report [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.963132] env[63293]: INFO nova.compute.manager [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Took 27.24 seconds to build instance. [ 639.014398] env[63293]: DEBUG nova.network.neutron [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.024816] env[63293]: DEBUG nova.compute.manager [req-4c15a36d-9a6f-4870-aadd-34f08b514528 req-c2f13533-5262-4b90-a126-7af685091c31 service nova] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Received event network-changed-453608e9-ae55-48dd-b58b-6189194d69c0 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.025018] env[63293]: DEBUG nova.compute.manager [req-4c15a36d-9a6f-4870-aadd-34f08b514528 req-c2f13533-5262-4b90-a126-7af685091c31 service nova] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Refreshing instance network info cache due to event network-changed-453608e9-ae55-48dd-b58b-6189194d69c0. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 639.025234] env[63293]: DEBUG oslo_concurrency.lockutils [req-4c15a36d-9a6f-4870-aadd-34f08b514528 req-c2f13533-5262-4b90-a126-7af685091c31 service nova] Acquiring lock "refresh_cache-eaaf2515-cc4b-42c7-aacd-14074c760924" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.073642] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "67839a84-c1a0-4728-88f4-47acf99f1fc4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.074029] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "67839a84-c1a0-4728-88f4-47acf99f1fc4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.104045] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "0df77939-d60c-4ccb-98f7-2325062ec76f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.104569] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "0df77939-d60c-4ccb-98f7-2325062ec76f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.429644] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.667s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.431009] env[63293]: DEBUG nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 639.433860] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.747s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.471027] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68e1e780-fde1-4968-9105-6374587aa084 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Lock "b2055187-edd8-4a8c-9811-183cf7ba3b39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.843s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.519320] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Releasing lock "refresh_cache-eaaf2515-cc4b-42c7-aacd-14074c760924" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.519320] env[63293]: DEBUG nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 639.519320] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 639.522216] env[63293]: DEBUG oslo_concurrency.lockutils [req-4c15a36d-9a6f-4870-aadd-34f08b514528 req-c2f13533-5262-4b90-a126-7af685091c31 service nova] Acquired lock "refresh_cache-eaaf2515-cc4b-42c7-aacd-14074c760924" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.522216] env[63293]: DEBUG nova.network.neutron [req-4c15a36d-9a6f-4870-aadd-34f08b514528 req-c2f13533-5262-4b90-a126-7af685091c31 service nova] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Refreshing network info cache for port 453608e9-ae55-48dd-b58b-6189194d69c0 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 639.523288] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c8ce8c1-8863-4480-998b-5845f3f4b247 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.540188] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd4b99f-a594-47a0-9446-1f315744b6c8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.569158] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eaaf2515-cc4b-42c7-aacd-14074c760924 could not be found. [ 639.569397] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.570138] env[63293]: INFO nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Took 0.05 seconds to destroy the instance on the hypervisor. [ 639.570138] env[63293]: DEBUG oslo.service.loopingcall [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.570576] env[63293]: DEBUG nova.compute.manager [-] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.570703] env[63293]: DEBUG nova.network.neutron [-] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.592238] env[63293]: DEBUG nova.network.neutron [-] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.728139] env[63293]: DEBUG nova.compute.manager [None req-1baadf4a-119f-4fa8-b042-19928c47f4ca tempest-ServerDiagnosticsV248Test-402368190 tempest-ServerDiagnosticsV248Test-402368190-project-admin] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 639.729683] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71567d6-c280-42ad-9450-b2bede5ead79 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.739380] env[63293]: INFO nova.compute.manager [None req-1baadf4a-119f-4fa8-b042-19928c47f4ca tempest-ServerDiagnosticsV248Test-402368190 tempest-ServerDiagnosticsV248Test-402368190-project-admin] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Retrieving diagnostics [ 639.740314] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8f1469-6a5d-4d0f-904e-027bd977f2fa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.939941] env[63293]: DEBUG nova.compute.utils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 639.947986] env[63293]: DEBUG nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 639.947986] env[63293]: DEBUG nova.network.neutron [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 639.974328] env[63293]: DEBUG nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 640.059923] env[63293]: DEBUG nova.policy [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc8b659f683a4a50b703c72413df89f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df458dc79fb54ba8b3f11a747da99969', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 640.063480] env[63293]: DEBUG nova.network.neutron [req-4c15a36d-9a6f-4870-aadd-34f08b514528 req-c2f13533-5262-4b90-a126-7af685091c31 service nova] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.096502] env[63293]: DEBUG nova.network.neutron [-] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.142314] env[63293]: ERROR nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 255fd7f1-f9a0-4469-80e0-980345a3cb6d, please check neutron logs for more information. [ 640.142314] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 640.142314] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.142314] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 640.142314] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.142314] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 640.142314] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.142314] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 640.142314] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.142314] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 640.142314] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.142314] env[63293]: ERROR nova.compute.manager raise self.value [ 640.142314] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.142314] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 640.142314] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.142314] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 640.142824] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.142824] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 640.142824] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 255fd7f1-f9a0-4469-80e0-980345a3cb6d, please check neutron logs for more information. [ 640.142824] env[63293]: ERROR nova.compute.manager [ 640.142824] env[63293]: Traceback (most recent call last): [ 640.142824] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 640.142824] env[63293]: listener.cb(fileno) [ 640.142824] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.142824] env[63293]: result = function(*args, **kwargs) [ 640.142824] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 640.142824] env[63293]: return func(*args, **kwargs) [ 640.142824] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.142824] env[63293]: raise e [ 640.142824] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.142824] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 640.142824] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.142824] env[63293]: created_port_ids = self._update_ports_for_instance( [ 640.142824] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.142824] env[63293]: with excutils.save_and_reraise_exception(): [ 640.142824] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.142824] env[63293]: self.force_reraise() [ 640.142824] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.142824] env[63293]: raise self.value [ 640.142824] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.142824] env[63293]: updated_port = self._update_port( [ 640.142824] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.142824] env[63293]: _ensure_no_port_binding_failure(port) [ 640.142824] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.142824] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 640.143553] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 255fd7f1-f9a0-4469-80e0-980345a3cb6d, please check neutron logs for more information. [ 640.143553] env[63293]: Removing descriptor: 18 [ 640.143553] env[63293]: ERROR nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 255fd7f1-f9a0-4469-80e0-980345a3cb6d, please check neutron logs for more information. [ 640.143553] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Traceback (most recent call last): [ 640.143553] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 640.143553] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] yield resources [ 640.143553] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.143553] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] self.driver.spawn(context, instance, image_meta, [ 640.143553] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 640.143553] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.143553] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.143553] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] vm_ref = self.build_virtual_machine(instance, [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] for vif in network_info: [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] return self._sync_wrapper(fn, *args, **kwargs) [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] self.wait() [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] self[:] = self._gt.wait() [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] return self._exit_event.wait() [ 640.144020] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] result = hub.switch() [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] return self.greenlet.switch() [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] result = function(*args, **kwargs) [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] return func(*args, **kwargs) [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] raise e [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] nwinfo = self.network_api.allocate_for_instance( [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.144346] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] created_port_ids = self._update_ports_for_instance( [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] with excutils.save_and_reraise_exception(): [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] self.force_reraise() [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] raise self.value [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] updated_port = self._update_port( [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] _ensure_no_port_binding_failure(port) [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.145329] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] raise exception.PortBindingFailed(port_id=port['id']) [ 640.145999] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] nova.exception.PortBindingFailed: Binding failed for port 255fd7f1-f9a0-4469-80e0-980345a3cb6d, please check neutron logs for more information. [ 640.145999] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] [ 640.145999] env[63293]: INFO nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Terminating instance [ 640.145999] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "refresh_cache-1814ebc5-7869-4ec6-8cf6-790daed5e271" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.145999] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "refresh_cache-1814ebc5-7869-4ec6-8cf6-790daed5e271" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.145999] env[63293]: DEBUG nova.network.neutron [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.227962] env[63293]: DEBUG nova.network.neutron [req-4c15a36d-9a6f-4870-aadd-34f08b514528 req-c2f13533-5262-4b90-a126-7af685091c31 service nova] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.451179] env[63293]: DEBUG nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 640.470373] env[63293]: DEBUG nova.network.neutron [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Successfully created port: 95e989b9-668a-43ee-8cdd-9069454e4e1c {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 640.506130] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759a6cb8-dab1-4042-bd54-e300bf4df84b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.514250] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.518576] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7b6626-f25a-418c-8356-7fff496e3c7d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.560611] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2acd1697-cd78-43fb-9726-b6ea90d45975 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.569028] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ef976e-8d29-4e01-b61e-6f97e78b40e5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.585635] env[63293]: DEBUG nova.compute.provider_tree [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.597874] env[63293]: INFO nova.compute.manager [-] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Took 1.03 seconds to deallocate network for instance. [ 640.601572] env[63293]: DEBUG nova.compute.claims [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.601888] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.672296] env[63293]: DEBUG nova.network.neutron [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.731269] env[63293]: DEBUG oslo_concurrency.lockutils [req-4c15a36d-9a6f-4870-aadd-34f08b514528 req-c2f13533-5262-4b90-a126-7af685091c31 service nova] Releasing lock "refresh_cache-eaaf2515-cc4b-42c7-aacd-14074c760924" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.731539] env[63293]: DEBUG nova.compute.manager [req-4c15a36d-9a6f-4870-aadd-34f08b514528 req-c2f13533-5262-4b90-a126-7af685091c31 service nova] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Received event network-vif-deleted-453608e9-ae55-48dd-b58b-6189194d69c0 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 640.794141] env[63293]: DEBUG nova.network.neutron [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.090563] env[63293]: DEBUG nova.scheduler.client.report [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.268788] env[63293]: DEBUG nova.compute.manager [req-8498e3e7-6f72-4bfe-b6b6-348bf410004d req-d9aed68d-6c2d-46d6-ba05-afe06f2f893a service nova] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Received event network-changed-255fd7f1-f9a0-4469-80e0-980345a3cb6d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.269234] env[63293]: DEBUG nova.compute.manager [req-8498e3e7-6f72-4bfe-b6b6-348bf410004d req-d9aed68d-6c2d-46d6-ba05-afe06f2f893a service nova] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Refreshing instance network info cache due to event network-changed-255fd7f1-f9a0-4469-80e0-980345a3cb6d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 641.269234] env[63293]: DEBUG oslo_concurrency.lockutils [req-8498e3e7-6f72-4bfe-b6b6-348bf410004d req-d9aed68d-6c2d-46d6-ba05-afe06f2f893a service nova] Acquiring lock "refresh_cache-1814ebc5-7869-4ec6-8cf6-790daed5e271" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.297442] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "refresh_cache-1814ebc5-7869-4ec6-8cf6-790daed5e271" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.297442] env[63293]: DEBUG nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 641.298686] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 641.299045] env[63293]: DEBUG oslo_concurrency.lockutils [req-8498e3e7-6f72-4bfe-b6b6-348bf410004d req-d9aed68d-6c2d-46d6-ba05-afe06f2f893a service nova] Acquired lock "refresh_cache-1814ebc5-7869-4ec6-8cf6-790daed5e271" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.299232] env[63293]: DEBUG nova.network.neutron [req-8498e3e7-6f72-4bfe-b6b6-348bf410004d req-d9aed68d-6c2d-46d6-ba05-afe06f2f893a service nova] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Refreshing network info cache for port 255fd7f1-f9a0-4469-80e0-980345a3cb6d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 641.300313] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b408fa63-efb8-4eb6-93a6-f76227b2efa6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.313069] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537425ed-714b-487b-ab42-c65c9b771149 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.343562] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1814ebc5-7869-4ec6-8cf6-790daed5e271 could not be found. [ 641.343949] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 641.344218] env[63293]: INFO nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Took 0.05 seconds to destroy the instance on the hypervisor. [ 641.344598] env[63293]: DEBUG oslo.service.loopingcall [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.345199] env[63293]: DEBUG nova.compute.manager [-] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.345393] env[63293]: DEBUG nova.network.neutron [-] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.362662] env[63293]: DEBUG nova.network.neutron [-] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.460201] env[63293]: DEBUG nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 641.482222] env[63293]: DEBUG nova.virt.hardware [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 641.482506] env[63293]: DEBUG nova.virt.hardware [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 641.482629] env[63293]: DEBUG nova.virt.hardware [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 641.482853] env[63293]: DEBUG nova.virt.hardware [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 641.486361] env[63293]: DEBUG nova.virt.hardware [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 641.486627] env[63293]: DEBUG nova.virt.hardware [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 641.487608] env[63293]: DEBUG nova.virt.hardware [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 641.487608] env[63293]: DEBUG nova.virt.hardware [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 641.487608] env[63293]: DEBUG nova.virt.hardware [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 641.487608] env[63293]: DEBUG nova.virt.hardware [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 641.487786] env[63293]: DEBUG nova.virt.hardware [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.492732] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51c1068-3d7a-4144-ab53-d1afee8a5897 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.501539] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a75bb6-c41a-47b1-9e4e-bb51042e9a78 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.601585] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.168s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.602786] env[63293]: ERROR nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 49767b23-0e55-492a-8ab0-6893c9580bda, please check neutron logs for more information. [ 641.602786] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Traceback (most recent call last): [ 641.602786] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.602786] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] self.driver.spawn(context, instance, image_meta, [ 641.602786] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 641.602786] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.602786] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.602786] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] vm_ref = self.build_virtual_machine(instance, [ 641.602786] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.602786] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.602786] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] for vif in network_info: [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] return self._sync_wrapper(fn, *args, **kwargs) [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] self.wait() [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] self[:] = self._gt.wait() [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] return self._exit_event.wait() [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] result = hub.switch() [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 641.603117] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] return self.greenlet.switch() [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] result = function(*args, **kwargs) [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] return func(*args, **kwargs) [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] raise e [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] nwinfo = self.network_api.allocate_for_instance( [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] created_port_ids = self._update_ports_for_instance( [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] with excutils.save_and_reraise_exception(): [ 641.603417] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.603728] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] self.force_reraise() [ 641.603728] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.603728] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] raise self.value [ 641.603728] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.603728] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] updated_port = self._update_port( [ 641.603728] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.603728] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] _ensure_no_port_binding_failure(port) [ 641.603728] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.603728] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] raise exception.PortBindingFailed(port_id=port['id']) [ 641.603728] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] nova.exception.PortBindingFailed: Binding failed for port 49767b23-0e55-492a-8ab0-6893c9580bda, please check neutron logs for more information. [ 641.603728] env[63293]: ERROR nova.compute.manager [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] [ 641.604417] env[63293]: DEBUG nova.compute.utils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Binding failed for port 49767b23-0e55-492a-8ab0-6893c9580bda, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 641.605954] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.829s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.608449] env[63293]: INFO nova.compute.claims [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 641.615097] env[63293]: DEBUG nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Build of instance a492b0a0-0bf1-4e64-bd51-ad131cd17a3b was re-scheduled: Binding failed for port 49767b23-0e55-492a-8ab0-6893c9580bda, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 641.615097] env[63293]: DEBUG nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 641.615097] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Acquiring lock "refresh_cache-a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.615097] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Acquired lock "refresh_cache-a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.615612] env[63293]: DEBUG nova.network.neutron [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.696580] env[63293]: ERROR nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 95e989b9-668a-43ee-8cdd-9069454e4e1c, please check neutron logs for more information. [ 641.696580] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 641.696580] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.696580] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 641.696580] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.696580] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 641.696580] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.696580] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 641.696580] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.696580] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 641.696580] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.696580] env[63293]: ERROR nova.compute.manager raise self.value [ 641.696580] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.696580] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 641.696580] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.696580] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 641.697041] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.697041] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 641.697041] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 95e989b9-668a-43ee-8cdd-9069454e4e1c, please check neutron logs for more information. [ 641.697041] env[63293]: ERROR nova.compute.manager [ 641.697041] env[63293]: Traceback (most recent call last): [ 641.697041] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 641.697041] env[63293]: listener.cb(fileno) [ 641.697041] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.697041] env[63293]: result = function(*args, **kwargs) [ 641.697041] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.697041] env[63293]: return func(*args, **kwargs) [ 641.697041] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.697041] env[63293]: raise e [ 641.697041] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.697041] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 641.697041] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.697041] env[63293]: created_port_ids = self._update_ports_for_instance( [ 641.697041] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.697041] env[63293]: with excutils.save_and_reraise_exception(): [ 641.697041] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.697041] env[63293]: self.force_reraise() [ 641.697041] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.697041] env[63293]: raise self.value [ 641.697041] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.697041] env[63293]: updated_port = self._update_port( [ 641.697041] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.697041] env[63293]: _ensure_no_port_binding_failure(port) [ 641.697041] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.697041] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 641.697903] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 95e989b9-668a-43ee-8cdd-9069454e4e1c, please check neutron logs for more information. [ 641.697903] env[63293]: Removing descriptor: 17 [ 641.697903] env[63293]: ERROR nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 95e989b9-668a-43ee-8cdd-9069454e4e1c, please check neutron logs for more information. [ 641.697903] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Traceback (most recent call last): [ 641.697903] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 641.697903] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] yield resources [ 641.697903] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.697903] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] self.driver.spawn(context, instance, image_meta, [ 641.697903] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 641.697903] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.697903] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.697903] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] vm_ref = self.build_virtual_machine(instance, [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] for vif in network_info: [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] return self._sync_wrapper(fn, *args, **kwargs) [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] self.wait() [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] self[:] = self._gt.wait() [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] return self._exit_event.wait() [ 641.698278] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] result = hub.switch() [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] return self.greenlet.switch() [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] result = function(*args, **kwargs) [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] return func(*args, **kwargs) [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] raise e [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] nwinfo = self.network_api.allocate_for_instance( [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.698635] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] created_port_ids = self._update_ports_for_instance( [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] with excutils.save_and_reraise_exception(): [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] self.force_reraise() [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] raise self.value [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] updated_port = self._update_port( [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] _ensure_no_port_binding_failure(port) [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.698985] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] raise exception.PortBindingFailed(port_id=port['id']) [ 641.699334] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] nova.exception.PortBindingFailed: Binding failed for port 95e989b9-668a-43ee-8cdd-9069454e4e1c, please check neutron logs for more information. [ 641.699334] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] [ 641.699334] env[63293]: INFO nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Terminating instance [ 641.701693] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Acquiring lock "refresh_cache-63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.701849] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Acquired lock "refresh_cache-63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.702032] env[63293]: DEBUG nova.network.neutron [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.830854] env[63293]: DEBUG nova.network.neutron [req-8498e3e7-6f72-4bfe-b6b6-348bf410004d req-d9aed68d-6c2d-46d6-ba05-afe06f2f893a service nova] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.840957] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Acquiring lock "b1b1ebe0-187b-4e23-9433-7dff420e3de3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.841210] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Lock "b1b1ebe0-187b-4e23-9433-7dff420e3de3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.865399] env[63293]: DEBUG nova.network.neutron [-] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.923148] env[63293]: DEBUG nova.network.neutron [req-8498e3e7-6f72-4bfe-b6b6-348bf410004d req-d9aed68d-6c2d-46d6-ba05-afe06f2f893a service nova] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.134952] env[63293]: DEBUG nova.network.neutron [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.204567] env[63293]: DEBUG nova.network.neutron [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.222056] env[63293]: DEBUG nova.network.neutron [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.296634] env[63293]: DEBUG nova.network.neutron [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.367676] env[63293]: INFO nova.compute.manager [-] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Took 1.02 seconds to deallocate network for instance. [ 642.372664] env[63293]: DEBUG nova.compute.claims [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 642.372834] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.427321] env[63293]: DEBUG oslo_concurrency.lockutils [req-8498e3e7-6f72-4bfe-b6b6-348bf410004d req-d9aed68d-6c2d-46d6-ba05-afe06f2f893a service nova] Releasing lock "refresh_cache-1814ebc5-7869-4ec6-8cf6-790daed5e271" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.427744] env[63293]: DEBUG nova.compute.manager [req-8498e3e7-6f72-4bfe-b6b6-348bf410004d req-d9aed68d-6c2d-46d6-ba05-afe06f2f893a service nova] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Received event network-vif-deleted-255fd7f1-f9a0-4469-80e0-980345a3cb6d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 642.712733] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Releasing lock "refresh_cache-a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.712733] env[63293]: DEBUG nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 642.712733] env[63293]: DEBUG nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.712733] env[63293]: DEBUG nova.network.neutron [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.740942] env[63293]: DEBUG nova.network.neutron [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.799863] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Releasing lock "refresh_cache-63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.800362] env[63293]: DEBUG nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 642.800537] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 642.801169] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a41760e6-0d1c-48c4-8a13-e9a4bfe388fe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.809637] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab701196-6f5f-413c-b2cb-2c041acd46d4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.835433] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5 could not be found. [ 642.835667] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 642.835850] env[63293]: INFO nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 642.836101] env[63293]: DEBUG oslo.service.loopingcall [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 642.839196] env[63293]: DEBUG nova.compute.manager [-] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.839340] env[63293]: DEBUG nova.network.neutron [-] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.854187] env[63293]: DEBUG nova.network.neutron [-] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.042705] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5ce80b-a0d2-4168-bce4-f5a8afd5d4a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.049843] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cadc9f1-dc3a-434d-8670-6fd93193aee1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.079870] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f34ad1-11d0-47f9-bfc4-9d3fd6375cd0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.087445] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f157295-ae58-4f50-b48e-89c7a9afe767 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.100833] env[63293]: DEBUG nova.compute.provider_tree [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.243173] env[63293]: DEBUG nova.network.neutron [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.293715] env[63293]: DEBUG nova.compute.manager [req-51d6a8dc-973e-4432-9bf0-9a3985898380 req-04e9f30e-ece4-498d-96fa-af8e6bde6211 service nova] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Received event network-changed-95e989b9-668a-43ee-8cdd-9069454e4e1c {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 643.293918] env[63293]: DEBUG nova.compute.manager [req-51d6a8dc-973e-4432-9bf0-9a3985898380 req-04e9f30e-ece4-498d-96fa-af8e6bde6211 service nova] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Refreshing instance network info cache due to event network-changed-95e989b9-668a-43ee-8cdd-9069454e4e1c. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 643.294145] env[63293]: DEBUG oslo_concurrency.lockutils [req-51d6a8dc-973e-4432-9bf0-9a3985898380 req-04e9f30e-ece4-498d-96fa-af8e6bde6211 service nova] Acquiring lock "refresh_cache-63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.294287] env[63293]: DEBUG oslo_concurrency.lockutils [req-51d6a8dc-973e-4432-9bf0-9a3985898380 req-04e9f30e-ece4-498d-96fa-af8e6bde6211 service nova] Acquired lock "refresh_cache-63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.294444] env[63293]: DEBUG nova.network.neutron [req-51d6a8dc-973e-4432-9bf0-9a3985898380 req-04e9f30e-ece4-498d-96fa-af8e6bde6211 service nova] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Refreshing network info cache for port 95e989b9-668a-43ee-8cdd-9069454e4e1c {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 643.359089] env[63293]: DEBUG nova.network.neutron [-] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.603641] env[63293]: DEBUG nova.scheduler.client.report [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.746640] env[63293]: INFO nova.compute.manager [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] [instance: a492b0a0-0bf1-4e64-bd51-ad131cd17a3b] Took 1.04 seconds to deallocate network for instance. [ 643.813137] env[63293]: DEBUG nova.network.neutron [req-51d6a8dc-973e-4432-9bf0-9a3985898380 req-04e9f30e-ece4-498d-96fa-af8e6bde6211 service nova] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.861813] env[63293]: INFO nova.compute.manager [-] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Took 1.02 seconds to deallocate network for instance. [ 643.865117] env[63293]: DEBUG nova.compute.claims [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 643.865117] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.918502] env[63293]: DEBUG nova.network.neutron [req-51d6a8dc-973e-4432-9bf0-9a3985898380 req-04e9f30e-ece4-498d-96fa-af8e6bde6211 service nova] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.109600] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.110210] env[63293]: DEBUG nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 644.113358] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.234s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.114792] env[63293]: INFO nova.compute.claims [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 644.421846] env[63293]: DEBUG oslo_concurrency.lockutils [req-51d6a8dc-973e-4432-9bf0-9a3985898380 req-04e9f30e-ece4-498d-96fa-af8e6bde6211 service nova] Releasing lock "refresh_cache-63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.421846] env[63293]: DEBUG nova.compute.manager [req-51d6a8dc-973e-4432-9bf0-9a3985898380 req-04e9f30e-ece4-498d-96fa-af8e6bde6211 service nova] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Received event network-vif-deleted-95e989b9-668a-43ee-8cdd-9069454e4e1c {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 644.618969] env[63293]: DEBUG nova.compute.utils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 644.622491] env[63293]: DEBUG nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 644.622664] env[63293]: DEBUG nova.network.neutron [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 644.685638] env[63293]: DEBUG nova.policy [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e67b9f3ab85401b9a7eb830334f2c33', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '775a99f9320a4a75a1e0d7804d645712', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 644.787842] env[63293]: INFO nova.scheduler.client.report [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Deleted allocations for instance a492b0a0-0bf1-4e64-bd51-ad131cd17a3b [ 645.126846] env[63293]: DEBUG nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 645.139787] env[63293]: DEBUG nova.network.neutron [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Successfully created port: 9f72114f-ae1c-4928-8d06-bc85d5cc0a96 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.302279] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad36188-5f71-4489-90bd-ce102bf25109 tempest-ServersWithSpecificFlavorTestJSON-434216921 tempest-ServersWithSpecificFlavorTestJSON-434216921-project-member] Lock "a492b0a0-0bf1-4e64-bd51-ad131cd17a3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.055s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.593655] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb3ba3a-f9b2-45ba-a612-2ad17e3d12f9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.601559] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c94de9-3656-4368-9c5d-e454d739e2f9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.637681] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d4736f-3324-4ba3-ab7c-d10bc2bebf4b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.645516] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872c239a-e87d-45c4-b922-eed31f44dcf0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.661836] env[63293]: DEBUG nova.compute.provider_tree [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.809462] env[63293]: DEBUG nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.142259] env[63293]: DEBUG nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 646.165716] env[63293]: DEBUG nova.scheduler.client.report [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.179515] env[63293]: DEBUG nova.virt.hardware [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 646.179909] env[63293]: DEBUG nova.virt.hardware [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 646.179909] env[63293]: DEBUG nova.virt.hardware [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 646.180117] env[63293]: DEBUG nova.virt.hardware [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 646.180270] env[63293]: DEBUG nova.virt.hardware [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 646.180434] env[63293]: DEBUG nova.virt.hardware [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 646.180721] env[63293]: DEBUG nova.virt.hardware [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 646.180920] env[63293]: DEBUG nova.virt.hardware [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 646.180975] env[63293]: DEBUG nova.virt.hardware [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 646.181192] env[63293]: DEBUG nova.virt.hardware [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 646.181381] env[63293]: DEBUG nova.virt.hardware [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.182233] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d8d583-5706-4f20-9ea0-9d478ff9e869 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.190570] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce665f8b-5b2c-4b54-a328-633f044c6a97 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.259173] env[63293]: DEBUG nova.compute.manager [req-a6dd1b6b-e759-4c73-93aa-7b0ae00c4952 req-e34e505c-db39-4882-b50b-c6a02dd0c4fb service nova] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Received event network-changed-9f72114f-ae1c-4928-8d06-bc85d5cc0a96 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 646.259353] env[63293]: DEBUG nova.compute.manager [req-a6dd1b6b-e759-4c73-93aa-7b0ae00c4952 req-e34e505c-db39-4882-b50b-c6a02dd0c4fb service nova] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Refreshing instance network info cache due to event network-changed-9f72114f-ae1c-4928-8d06-bc85d5cc0a96. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 646.259593] env[63293]: DEBUG oslo_concurrency.lockutils [req-a6dd1b6b-e759-4c73-93aa-7b0ae00c4952 req-e34e505c-db39-4882-b50b-c6a02dd0c4fb service nova] Acquiring lock "refresh_cache-87c7bb2c-46ff-45f7-8967-14192f2b42e4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.259769] env[63293]: DEBUG oslo_concurrency.lockutils [req-a6dd1b6b-e759-4c73-93aa-7b0ae00c4952 req-e34e505c-db39-4882-b50b-c6a02dd0c4fb service nova] Acquired lock "refresh_cache-87c7bb2c-46ff-45f7-8967-14192f2b42e4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.259946] env[63293]: DEBUG nova.network.neutron [req-a6dd1b6b-e759-4c73-93aa-7b0ae00c4952 req-e34e505c-db39-4882-b50b-c6a02dd0c4fb service nova] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Refreshing network info cache for port 9f72114f-ae1c-4928-8d06-bc85d5cc0a96 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 646.352875] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.432488] env[63293]: ERROR nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9f72114f-ae1c-4928-8d06-bc85d5cc0a96, please check neutron logs for more information. [ 646.432488] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 646.432488] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.432488] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 646.432488] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.432488] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 646.432488] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.432488] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 646.432488] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.432488] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 646.432488] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.432488] env[63293]: ERROR nova.compute.manager raise self.value [ 646.432488] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.432488] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 646.432488] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.432488] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 646.433021] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.433021] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 646.433021] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9f72114f-ae1c-4928-8d06-bc85d5cc0a96, please check neutron logs for more information. [ 646.433021] env[63293]: ERROR nova.compute.manager [ 646.433021] env[63293]: Traceback (most recent call last): [ 646.433021] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 646.433021] env[63293]: listener.cb(fileno) [ 646.433021] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.433021] env[63293]: result = function(*args, **kwargs) [ 646.433021] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.433021] env[63293]: return func(*args, **kwargs) [ 646.433021] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.433021] env[63293]: raise e [ 646.433021] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.433021] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 646.433021] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.433021] env[63293]: created_port_ids = self._update_ports_for_instance( [ 646.433021] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.433021] env[63293]: with excutils.save_and_reraise_exception(): [ 646.433021] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.433021] env[63293]: self.force_reraise() [ 646.433021] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.433021] env[63293]: raise self.value [ 646.433021] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.433021] env[63293]: updated_port = self._update_port( [ 646.433021] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.433021] env[63293]: _ensure_no_port_binding_failure(port) [ 646.433021] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.433021] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 646.433888] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 9f72114f-ae1c-4928-8d06-bc85d5cc0a96, please check neutron logs for more information. [ 646.433888] env[63293]: Removing descriptor: 17 [ 646.433888] env[63293]: ERROR nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9f72114f-ae1c-4928-8d06-bc85d5cc0a96, please check neutron logs for more information. [ 646.433888] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Traceback (most recent call last): [ 646.433888] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 646.433888] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] yield resources [ 646.433888] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.433888] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] self.driver.spawn(context, instance, image_meta, [ 646.433888] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 646.433888] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.433888] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.433888] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] vm_ref = self.build_virtual_machine(instance, [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] for vif in network_info: [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] return self._sync_wrapper(fn, *args, **kwargs) [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] self.wait() [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] self[:] = self._gt.wait() [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] return self._exit_event.wait() [ 646.434493] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] result = hub.switch() [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] return self.greenlet.switch() [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] result = function(*args, **kwargs) [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] return func(*args, **kwargs) [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] raise e [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] nwinfo = self.network_api.allocate_for_instance( [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.434990] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] created_port_ids = self._update_ports_for_instance( [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] with excutils.save_and_reraise_exception(): [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] self.force_reraise() [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] raise self.value [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] updated_port = self._update_port( [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] _ensure_no_port_binding_failure(port) [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.435377] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] raise exception.PortBindingFailed(port_id=port['id']) [ 646.435844] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] nova.exception.PortBindingFailed: Binding failed for port 9f72114f-ae1c-4928-8d06-bc85d5cc0a96, please check neutron logs for more information. [ 646.435844] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] [ 646.435844] env[63293]: INFO nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Terminating instance [ 646.435982] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquiring lock "refresh_cache-87c7bb2c-46ff-45f7-8967-14192f2b42e4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.669777] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.670340] env[63293]: DEBUG nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 646.675853] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.479s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.676740] env[63293]: INFO nova.compute.claims [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.781561] env[63293]: DEBUG nova.network.neutron [req-a6dd1b6b-e759-4c73-93aa-7b0ae00c4952 req-e34e505c-db39-4882-b50b-c6a02dd0c4fb service nova] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.870671] env[63293]: DEBUG nova.network.neutron [req-a6dd1b6b-e759-4c73-93aa-7b0ae00c4952 req-e34e505c-db39-4882-b50b-c6a02dd0c4fb service nova] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.178817] env[63293]: DEBUG nova.compute.utils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 647.178817] env[63293]: DEBUG nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 647.178817] env[63293]: DEBUG nova.network.neutron [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 647.230911] env[63293]: DEBUG nova.policy [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a31e6add8ee54ce69bc3ffd9a1056f1a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'db20579b85394de3bfb7afdd7dd78207', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 647.373834] env[63293]: DEBUG oslo_concurrency.lockutils [req-a6dd1b6b-e759-4c73-93aa-7b0ae00c4952 req-e34e505c-db39-4882-b50b-c6a02dd0c4fb service nova] Releasing lock "refresh_cache-87c7bb2c-46ff-45f7-8967-14192f2b42e4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.377956] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquired lock "refresh_cache-87c7bb2c-46ff-45f7-8967-14192f2b42e4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.377956] env[63293]: DEBUG nova.network.neutron [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.653952] env[63293]: DEBUG nova.network.neutron [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Successfully created port: 89a49c2d-7349-467b-96ab-5b38bc334f87 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 647.681721] env[63293]: DEBUG nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 647.913717] env[63293]: DEBUG nova.network.neutron [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.035132] env[63293]: DEBUG nova.network.neutron [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.249135] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bdea539-28ee-4ce1-bb95-69790a9dc49a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.256824] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bca1e3f-f1f4-4b44-9794-2da3ad224d4e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.297577] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3831b492-c7e4-41df-8e4f-e2f114de965e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.302365] env[63293]: DEBUG nova.compute.manager [req-1ac68eeb-2d13-48d4-951e-1f4e8c8e4501 req-289022c7-893d-48cb-8c0f-3a659aa09c3e service nova] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Received event network-vif-deleted-9f72114f-ae1c-4928-8d06-bc85d5cc0a96 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.306969] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1bac498-a1d7-4771-a565-1b01f3149584 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.322110] env[63293]: DEBUG nova.compute.provider_tree [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.540482] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Releasing lock "refresh_cache-87c7bb2c-46ff-45f7-8967-14192f2b42e4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.541097] env[63293]: DEBUG nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 648.541293] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.541596] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1dfdfa3b-57cb-4c52-93ef-fb06a2687b53 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.553357] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ba9918-424a-4d57-aa19-ddd9b573bfc3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.581530] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 87c7bb2c-46ff-45f7-8967-14192f2b42e4 could not be found. [ 648.581748] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 648.581931] env[63293]: INFO nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 648.582193] env[63293]: DEBUG oslo.service.loopingcall [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.582389] env[63293]: DEBUG nova.compute.manager [-] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.582483] env[63293]: DEBUG nova.network.neutron [-] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 648.608151] env[63293]: DEBUG nova.network.neutron [-] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.697538] env[63293]: DEBUG nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 648.724045] env[63293]: DEBUG nova.virt.hardware [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 648.724277] env[63293]: DEBUG nova.virt.hardware [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 648.724433] env[63293]: DEBUG nova.virt.hardware [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 648.724617] env[63293]: DEBUG nova.virt.hardware [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 648.724781] env[63293]: DEBUG nova.virt.hardware [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 648.724903] env[63293]: DEBUG nova.virt.hardware [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 648.725369] env[63293]: DEBUG nova.virt.hardware [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 648.725573] env[63293]: DEBUG nova.virt.hardware [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 648.725745] env[63293]: DEBUG nova.virt.hardware [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 648.725905] env[63293]: DEBUG nova.virt.hardware [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 648.726316] env[63293]: DEBUG nova.virt.hardware [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 648.727285] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb628fa-3a51-4cde-a2e0-e4feffbb2a67 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.735012] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01cff39f-847d-456a-96cc-dabd5194a586 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.825671] env[63293]: DEBUG nova.scheduler.client.report [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.858468] env[63293]: ERROR nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 89a49c2d-7349-467b-96ab-5b38bc334f87, please check neutron logs for more information. [ 648.858468] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.858468] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.858468] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.858468] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.858468] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.858468] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.858468] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.858468] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.858468] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 648.858468] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.858468] env[63293]: ERROR nova.compute.manager raise self.value [ 648.858468] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.858468] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.858468] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.858468] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.858900] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.858900] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.858900] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 89a49c2d-7349-467b-96ab-5b38bc334f87, please check neutron logs for more information. [ 648.858900] env[63293]: ERROR nova.compute.manager [ 648.858900] env[63293]: Traceback (most recent call last): [ 648.858900] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.858900] env[63293]: listener.cb(fileno) [ 648.858900] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.858900] env[63293]: result = function(*args, **kwargs) [ 648.858900] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 648.858900] env[63293]: return func(*args, **kwargs) [ 648.858900] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.858900] env[63293]: raise e [ 648.858900] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.858900] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 648.858900] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.858900] env[63293]: created_port_ids = self._update_ports_for_instance( [ 648.858900] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.858900] env[63293]: with excutils.save_and_reraise_exception(): [ 648.858900] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.858900] env[63293]: self.force_reraise() [ 648.858900] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.858900] env[63293]: raise self.value [ 648.858900] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.858900] env[63293]: updated_port = self._update_port( [ 648.858900] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.858900] env[63293]: _ensure_no_port_binding_failure(port) [ 648.858900] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.858900] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.859653] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 89a49c2d-7349-467b-96ab-5b38bc334f87, please check neutron logs for more information. [ 648.859653] env[63293]: Removing descriptor: 17 [ 648.861384] env[63293]: ERROR nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 89a49c2d-7349-467b-96ab-5b38bc334f87, please check neutron logs for more information. [ 648.861384] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Traceback (most recent call last): [ 648.861384] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 648.861384] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] yield resources [ 648.861384] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.861384] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] self.driver.spawn(context, instance, image_meta, [ 648.861384] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 648.861384] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.861384] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.861384] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] vm_ref = self.build_virtual_machine(instance, [ 648.861384] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] for vif in network_info: [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] return self._sync_wrapper(fn, *args, **kwargs) [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] self.wait() [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] self[:] = self._gt.wait() [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] return self._exit_event.wait() [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.861747] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] result = hub.switch() [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] return self.greenlet.switch() [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] result = function(*args, **kwargs) [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] return func(*args, **kwargs) [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] raise e [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] nwinfo = self.network_api.allocate_for_instance( [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] created_port_ids = self._update_ports_for_instance( [ 648.862097] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] with excutils.save_and_reraise_exception(): [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] self.force_reraise() [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] raise self.value [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] updated_port = self._update_port( [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] _ensure_no_port_binding_failure(port) [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] raise exception.PortBindingFailed(port_id=port['id']) [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] nova.exception.PortBindingFailed: Binding failed for port 89a49c2d-7349-467b-96ab-5b38bc334f87, please check neutron logs for more information. [ 648.862428] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] [ 648.862758] env[63293]: INFO nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Terminating instance [ 648.863193] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquiring lock "refresh_cache-769c2643-57eb-456c-8083-997a8c9c48cc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.863353] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquired lock "refresh_cache-769c2643-57eb-456c-8083-997a8c9c48cc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.863515] env[63293]: DEBUG nova.network.neutron [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.111015] env[63293]: DEBUG nova.network.neutron [-] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.331803] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.656s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.332208] env[63293]: DEBUG nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 649.335549] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.192s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.340544] env[63293]: INFO nova.compute.claims [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.420474] env[63293]: DEBUG nova.network.neutron [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.515742] env[63293]: DEBUG nova.network.neutron [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.613990] env[63293]: INFO nova.compute.manager [-] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Took 1.03 seconds to deallocate network for instance. [ 649.617217] env[63293]: DEBUG nova.compute.claims [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 649.617395] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.847922] env[63293]: DEBUG nova.compute.utils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 649.853493] env[63293]: DEBUG nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 649.853493] env[63293]: DEBUG nova.network.neutron [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 649.932698] env[63293]: DEBUG nova.policy [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e67b9f3ab85401b9a7eb830334f2c33', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '775a99f9320a4a75a1e0d7804d645712', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 650.022724] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Releasing lock "refresh_cache-769c2643-57eb-456c-8083-997a8c9c48cc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.022724] env[63293]: DEBUG nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 650.022724] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 650.022724] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-edf7a04f-9c47-46f4-9f94-820f6cf154a8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.032564] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc595dfb-62dc-4747-a821-3f9a7c7065b6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.063015] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 769c2643-57eb-456c-8083-997a8c9c48cc could not be found. [ 650.063015] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 650.063015] env[63293]: INFO nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 650.063015] env[63293]: DEBUG oslo.service.loopingcall [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.063015] env[63293]: DEBUG nova.compute.manager [-] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.063015] env[63293]: DEBUG nova.network.neutron [-] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 650.089140] env[63293]: DEBUG nova.network.neutron [-] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.276054] env[63293]: DEBUG nova.network.neutron [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Successfully created port: 52e67338-7b66-47ff-8bea-0783bb5999a1 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 650.315589] env[63293]: DEBUG nova.compute.manager [None req-bbe489f3-a3d0-4d66-9012-3831a16fbae6 tempest-ServerDiagnosticsV248Test-402368190 tempest-ServerDiagnosticsV248Test-402368190-project-admin] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 650.318823] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b30505-d082-4769-890b-aed0b3240624 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.327614] env[63293]: INFO nova.compute.manager [None req-bbe489f3-a3d0-4d66-9012-3831a16fbae6 tempest-ServerDiagnosticsV248Test-402368190 tempest-ServerDiagnosticsV248Test-402368190-project-admin] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Retrieving diagnostics [ 650.328386] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0962446e-0b98-4b4c-88ca-e1f45b7c3286 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.364668] env[63293]: DEBUG nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 650.373737] env[63293]: DEBUG nova.compute.manager [req-075408f8-f00e-4b68-ae8d-7f406ff0fe6b req-318d7f96-9120-4a64-84fc-e6d6017e1c3e service nova] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Received event network-changed-89a49c2d-7349-467b-96ab-5b38bc334f87 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.373941] env[63293]: DEBUG nova.compute.manager [req-075408f8-f00e-4b68-ae8d-7f406ff0fe6b req-318d7f96-9120-4a64-84fc-e6d6017e1c3e service nova] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Refreshing instance network info cache due to event network-changed-89a49c2d-7349-467b-96ab-5b38bc334f87. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 650.374177] env[63293]: DEBUG oslo_concurrency.lockutils [req-075408f8-f00e-4b68-ae8d-7f406ff0fe6b req-318d7f96-9120-4a64-84fc-e6d6017e1c3e service nova] Acquiring lock "refresh_cache-769c2643-57eb-456c-8083-997a8c9c48cc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.374316] env[63293]: DEBUG oslo_concurrency.lockutils [req-075408f8-f00e-4b68-ae8d-7f406ff0fe6b req-318d7f96-9120-4a64-84fc-e6d6017e1c3e service nova] Acquired lock "refresh_cache-769c2643-57eb-456c-8083-997a8c9c48cc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.374718] env[63293]: DEBUG nova.network.neutron [req-075408f8-f00e-4b68-ae8d-7f406ff0fe6b req-318d7f96-9120-4a64-84fc-e6d6017e1c3e service nova] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Refreshing network info cache for port 89a49c2d-7349-467b-96ab-5b38bc334f87 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 650.592762] env[63293]: DEBUG nova.network.neutron [-] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.870966] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c294ea9-60b4-4c46-b375-0e19f3597f7d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.884480] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b573811-f730-487a-b4ef-0742df54f17f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.918483] env[63293]: DEBUG nova.network.neutron [req-075408f8-f00e-4b68-ae8d-7f406ff0fe6b req-318d7f96-9120-4a64-84fc-e6d6017e1c3e service nova] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.920519] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e06258-4c2c-4286-aaa8-5f13bc4ac0a8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.931016] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4599531a-3e87-4573-8cbd-3e37e26e7a66 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.946134] env[63293]: DEBUG nova.compute.provider_tree [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.994108] env[63293]: DEBUG nova.network.neutron [req-075408f8-f00e-4b68-ae8d-7f406ff0fe6b req-318d7f96-9120-4a64-84fc-e6d6017e1c3e service nova] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.095321] env[63293]: INFO nova.compute.manager [-] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Took 1.03 seconds to deallocate network for instance. [ 651.098976] env[63293]: DEBUG nova.compute.claims [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 651.099182] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.383786] env[63293]: DEBUG nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.420173] env[63293]: DEBUG nova.virt.hardware [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.420173] env[63293]: DEBUG nova.virt.hardware [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.420173] env[63293]: DEBUG nova.virt.hardware [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.420331] env[63293]: DEBUG nova.virt.hardware [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.420331] env[63293]: DEBUG nova.virt.hardware [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.420331] env[63293]: DEBUG nova.virt.hardware [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.420331] env[63293]: DEBUG nova.virt.hardware [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.420331] env[63293]: DEBUG nova.virt.hardware [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.420461] env[63293]: DEBUG nova.virt.hardware [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.420461] env[63293]: DEBUG nova.virt.hardware [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.420461] env[63293]: DEBUG nova.virt.hardware [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.420594] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775854d2-7473-4519-b656-3d6bb343a826 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.430251] env[63293]: ERROR nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52e67338-7b66-47ff-8bea-0783bb5999a1, please check neutron logs for more information. [ 651.430251] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.430251] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.430251] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.430251] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.430251] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.430251] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.430251] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.430251] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.430251] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 651.430251] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.430251] env[63293]: ERROR nova.compute.manager raise self.value [ 651.430251] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.430251] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.430251] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.430251] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.430805] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.430805] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.430805] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52e67338-7b66-47ff-8bea-0783bb5999a1, please check neutron logs for more information. [ 651.430805] env[63293]: ERROR nova.compute.manager [ 651.430805] env[63293]: Traceback (most recent call last): [ 651.430805] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.430805] env[63293]: listener.cb(fileno) [ 651.430805] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.430805] env[63293]: result = function(*args, **kwargs) [ 651.430805] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.430805] env[63293]: return func(*args, **kwargs) [ 651.430805] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.430805] env[63293]: raise e [ 651.430805] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.430805] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 651.430805] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.430805] env[63293]: created_port_ids = self._update_ports_for_instance( [ 651.430805] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.430805] env[63293]: with excutils.save_and_reraise_exception(): [ 651.430805] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.430805] env[63293]: self.force_reraise() [ 651.430805] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.430805] env[63293]: raise self.value [ 651.430805] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.430805] env[63293]: updated_port = self._update_port( [ 651.430805] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.430805] env[63293]: _ensure_no_port_binding_failure(port) [ 651.430805] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.430805] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.431502] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 52e67338-7b66-47ff-8bea-0783bb5999a1, please check neutron logs for more information. [ 651.431502] env[63293]: Removing descriptor: 17 [ 651.431733] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8be54db-5c22-4ad1-a2d0-0c1dcea51025 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.447139] env[63293]: ERROR nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52e67338-7b66-47ff-8bea-0783bb5999a1, please check neutron logs for more information. [ 651.447139] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Traceback (most recent call last): [ 651.447139] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.447139] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] yield resources [ 651.447139] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.447139] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] self.driver.spawn(context, instance, image_meta, [ 651.447139] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 651.447139] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.447139] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.447139] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] vm_ref = self.build_virtual_machine(instance, [ 651.447139] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] for vif in network_info: [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] return self._sync_wrapper(fn, *args, **kwargs) [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] self.wait() [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] self[:] = self._gt.wait() [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] return self._exit_event.wait() [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 651.447438] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] current.throw(*self._exc) [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] result = function(*args, **kwargs) [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] return func(*args, **kwargs) [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] raise e [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] nwinfo = self.network_api.allocate_for_instance( [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] created_port_ids = self._update_ports_for_instance( [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] with excutils.save_and_reraise_exception(): [ 651.448038] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.448536] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] self.force_reraise() [ 651.448536] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.448536] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] raise self.value [ 651.448536] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.448536] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] updated_port = self._update_port( [ 651.448536] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.448536] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] _ensure_no_port_binding_failure(port) [ 651.448536] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.448536] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] raise exception.PortBindingFailed(port_id=port['id']) [ 651.448536] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] nova.exception.PortBindingFailed: Binding failed for port 52e67338-7b66-47ff-8bea-0783bb5999a1, please check neutron logs for more information. [ 651.448536] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] [ 651.448536] env[63293]: INFO nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Terminating instance [ 651.449490] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquiring lock "refresh_cache-b2c1e806-d0ae-455f-95e9-405a9557697f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.449654] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquired lock "refresh_cache-b2c1e806-d0ae-455f-95e9-405a9557697f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.449958] env[63293]: DEBUG nova.network.neutron [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.451297] env[63293]: DEBUG nova.scheduler.client.report [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.500291] env[63293]: DEBUG oslo_concurrency.lockutils [req-075408f8-f00e-4b68-ae8d-7f406ff0fe6b req-318d7f96-9120-4a64-84fc-e6d6017e1c3e service nova] Releasing lock "refresh_cache-769c2643-57eb-456c-8083-997a8c9c48cc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.500569] env[63293]: DEBUG nova.compute.manager [req-075408f8-f00e-4b68-ae8d-7f406ff0fe6b req-318d7f96-9120-4a64-84fc-e6d6017e1c3e service nova] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Received event network-vif-deleted-89a49c2d-7349-467b-96ab-5b38bc334f87 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.611725] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquiring lock "b2055187-edd8-4a8c-9811-183cf7ba3b39" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.611725] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Lock "b2055187-edd8-4a8c-9811-183cf7ba3b39" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.611725] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquiring lock "b2055187-edd8-4a8c-9811-183cf7ba3b39-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.611725] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Lock "b2055187-edd8-4a8c-9811-183cf7ba3b39-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.611855] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Lock "b2055187-edd8-4a8c-9811-183cf7ba3b39-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.617028] env[63293]: INFO nova.compute.manager [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Terminating instance [ 651.618359] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquiring lock "refresh_cache-b2055187-edd8-4a8c-9811-183cf7ba3b39" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.618359] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquired lock "refresh_cache-b2055187-edd8-4a8c-9811-183cf7ba3b39" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.618359] env[63293]: DEBUG nova.network.neutron [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.961485] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.626s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.962124] env[63293]: DEBUG nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 651.965342] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.019s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.965524] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.965737] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63293) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 651.966092] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.612s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.973017] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc0ec902-8aed-4967-9027-8bc984d5a333 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.979296] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96671288-f654-4fb5-9445-c3abb05bdab2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.003128] env[63293]: DEBUG nova.network.neutron [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.005530] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8df09b2-408c-4c58-9a69-6af1620e10da {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.013410] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d4a5f9-3972-4319-bb98-161221239b35 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.046801] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181517MB free_disk=132GB free_vcpus=48 pci_devices=None {{(pid=63293) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 652.047230] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.088176] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Acquiring lock "8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.089198] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Lock "8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.128649] env[63293]: DEBUG nova.network.neutron [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.138158] env[63293]: DEBUG nova.network.neutron [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.197697] env[63293]: DEBUG nova.network.neutron [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.399100] env[63293]: DEBUG nova.compute.manager [req-c0803f35-b813-43df-b37c-22498b604927 req-0bbb0609-5d22-4d34-9828-7107d785b852 service nova] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Received event network-changed-52e67338-7b66-47ff-8bea-0783bb5999a1 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.399358] env[63293]: DEBUG nova.compute.manager [req-c0803f35-b813-43df-b37c-22498b604927 req-0bbb0609-5d22-4d34-9828-7107d785b852 service nova] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Refreshing instance network info cache due to event network-changed-52e67338-7b66-47ff-8bea-0783bb5999a1. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 652.399516] env[63293]: DEBUG oslo_concurrency.lockutils [req-c0803f35-b813-43df-b37c-22498b604927 req-0bbb0609-5d22-4d34-9828-7107d785b852 service nova] Acquiring lock "refresh_cache-b2c1e806-d0ae-455f-95e9-405a9557697f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.475291] env[63293]: DEBUG nova.compute.utils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.476884] env[63293]: DEBUG nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 652.477076] env[63293]: DEBUG nova.network.neutron [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 652.560559] env[63293]: DEBUG nova.policy [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2312d48038d7474481a0709267645780', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '866b347100454019b07f63922b995bb6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 652.631073] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Releasing lock "refresh_cache-b2c1e806-d0ae-455f-95e9-405a9557697f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.631498] env[63293]: DEBUG nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.631808] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.632205] env[63293]: DEBUG oslo_concurrency.lockutils [req-c0803f35-b813-43df-b37c-22498b604927 req-0bbb0609-5d22-4d34-9828-7107d785b852 service nova] Acquired lock "refresh_cache-b2c1e806-d0ae-455f-95e9-405a9557697f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.632205] env[63293]: DEBUG nova.network.neutron [req-c0803f35-b813-43df-b37c-22498b604927 req-0bbb0609-5d22-4d34-9828-7107d785b852 service nova] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Refreshing network info cache for port 52e67338-7b66-47ff-8bea-0783bb5999a1 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 652.633242] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10e60772-d86e-4310-b352-0983aedb272e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.644731] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3055e46-dbd1-4c81-a2a0-20472f71a3bf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.675174] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b2c1e806-d0ae-455f-95e9-405a9557697f could not be found. [ 652.675321] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.675517] env[63293]: INFO nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 652.675800] env[63293]: DEBUG oslo.service.loopingcall [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.676643] env[63293]: DEBUG nova.compute.manager [-] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.676744] env[63293]: DEBUG nova.network.neutron [-] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.700482] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Releasing lock "refresh_cache-b2055187-edd8-4a8c-9811-183cf7ba3b39" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.700928] env[63293]: DEBUG nova.compute.manager [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.701268] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.702324] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf9e52d-3583-4d56-a558-33ebc195998b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.705562] env[63293]: DEBUG nova.network.neutron [-] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.712683] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 652.713055] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11d06a41-0d14-4197-b94d-17f469753807 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.720115] env[63293]: DEBUG oslo_vmware.api [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for the task: (returnval){ [ 652.720115] env[63293]: value = "task-1327461" [ 652.720115] env[63293]: _type = "Task" [ 652.720115] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.733707] env[63293]: DEBUG oslo_vmware.api [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327461, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.979932] env[63293]: DEBUG nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 652.998790] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f5b8c9-f168-4bd3-88fc-6b03c5a39b34 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.006757] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fdb9bd-b452-4b7c-a0e2-f42643e59639 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.036950] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffb8a2a-de91-4cd1-9858-aff4fdc1ce51 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.044817] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5930d19-9336-4582-85a3-c4594e3059a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.061097] env[63293]: DEBUG nova.compute.provider_tree [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.152419] env[63293]: DEBUG nova.network.neutron [req-c0803f35-b813-43df-b37c-22498b604927 req-0bbb0609-5d22-4d34-9828-7107d785b852 service nova] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.208777] env[63293]: DEBUG nova.network.neutron [-] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.220265] env[63293]: DEBUG nova.network.neutron [req-c0803f35-b813-43df-b37c-22498b604927 req-0bbb0609-5d22-4d34-9828-7107d785b852 service nova] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.231804] env[63293]: DEBUG oslo_vmware.api [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327461, 'name': PowerOffVM_Task, 'duration_secs': 0.107576} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.232834] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 653.233315] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 653.233582] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e35bc641-49d2-40fd-95c7-6d7dbde99bbf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.256800] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 653.257016] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 653.258018] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Deleting the datastore file [datastore1] b2055187-edd8-4a8c-9811-183cf7ba3b39 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 653.258018] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ecbcab8-3c2e-4c12-866d-4b6dc5a30c30 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.264344] env[63293]: DEBUG oslo_vmware.api [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for the task: (returnval){ [ 653.264344] env[63293]: value = "task-1327463" [ 653.264344] env[63293]: _type = "Task" [ 653.264344] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.272670] env[63293]: DEBUG oslo_vmware.api [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327463, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.410496] env[63293]: DEBUG nova.network.neutron [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Successfully created port: 6683eea7-366f-4165-846e-f1103c0a0264 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.563309] env[63293]: DEBUG nova.scheduler.client.report [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.711961] env[63293]: INFO nova.compute.manager [-] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Took 1.04 seconds to deallocate network for instance. [ 653.714510] env[63293]: DEBUG nova.compute.claims [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 653.714690] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.726721] env[63293]: DEBUG oslo_concurrency.lockutils [req-c0803f35-b813-43df-b37c-22498b604927 req-0bbb0609-5d22-4d34-9828-7107d785b852 service nova] Releasing lock "refresh_cache-b2c1e806-d0ae-455f-95e9-405a9557697f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.726959] env[63293]: DEBUG nova.compute.manager [req-c0803f35-b813-43df-b37c-22498b604927 req-0bbb0609-5d22-4d34-9828-7107d785b852 service nova] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Received event network-vif-deleted-52e67338-7b66-47ff-8bea-0783bb5999a1 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.773416] env[63293]: DEBUG oslo_vmware.api [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Task: {'id': task-1327463, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100657} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.773718] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 653.773906] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 653.774101] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.774896] env[63293]: INFO nova.compute.manager [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Took 1.07 seconds to destroy the instance on the hypervisor. [ 653.774896] env[63293]: DEBUG oslo.service.loopingcall [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.774896] env[63293]: DEBUG nova.compute.manager [-] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.774896] env[63293]: DEBUG nova.network.neutron [-] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.805842] env[63293]: DEBUG nova.network.neutron [-] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.990282] env[63293]: DEBUG nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 654.017583] env[63293]: DEBUG nova.virt.hardware [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 654.017723] env[63293]: DEBUG nova.virt.hardware [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 654.017881] env[63293]: DEBUG nova.virt.hardware [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 654.018077] env[63293]: DEBUG nova.virt.hardware [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 654.018261] env[63293]: DEBUG nova.virt.hardware [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 654.018369] env[63293]: DEBUG nova.virt.hardware [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 654.018569] env[63293]: DEBUG nova.virt.hardware [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 654.018744] env[63293]: DEBUG nova.virt.hardware [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 654.018904] env[63293]: DEBUG nova.virt.hardware [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 654.019075] env[63293]: DEBUG nova.virt.hardware [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 654.019249] env[63293]: DEBUG nova.virt.hardware [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 654.020120] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b00b2d-4476-454f-8055-2fa7f54118c2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.028879] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70460ed7-e796-4825-a838-b4150d9b40ea {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.070043] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.104s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.070857] env[63293]: ERROR nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4993189a-a876-4140-8985-d263769da28b, please check neutron logs for more information. [ 654.070857] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Traceback (most recent call last): [ 654.070857] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.070857] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] self.driver.spawn(context, instance, image_meta, [ 654.070857] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 654.070857] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.070857] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.070857] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] vm_ref = self.build_virtual_machine(instance, [ 654.070857] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.070857] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.070857] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] for vif in network_info: [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] return self._sync_wrapper(fn, *args, **kwargs) [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] self.wait() [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] self[:] = self._gt.wait() [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] return self._exit_event.wait() [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] result = hub.switch() [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.071188] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] return self.greenlet.switch() [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] result = function(*args, **kwargs) [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] return func(*args, **kwargs) [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] raise e [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] nwinfo = self.network_api.allocate_for_instance( [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] created_port_ids = self._update_ports_for_instance( [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] with excutils.save_and_reraise_exception(): [ 654.071489] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.071817] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] self.force_reraise() [ 654.071817] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.071817] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] raise self.value [ 654.071817] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.071817] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] updated_port = self._update_port( [ 654.071817] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.071817] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] _ensure_no_port_binding_failure(port) [ 654.071817] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.071817] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] raise exception.PortBindingFailed(port_id=port['id']) [ 654.071817] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] nova.exception.PortBindingFailed: Binding failed for port 4993189a-a876-4140-8985-d263769da28b, please check neutron logs for more information. [ 654.071817] env[63293]: ERROR nova.compute.manager [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] [ 654.072107] env[63293]: DEBUG nova.compute.utils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Binding failed for port 4993189a-a876-4140-8985-d263769da28b, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 654.073114] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.559s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.074872] env[63293]: INFO nova.compute.claims [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 654.078404] env[63293]: DEBUG nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Build of instance 510b9ea6-04f1-4704-b5e4-e19db57769a8 was re-scheduled: Binding failed for port 4993189a-a876-4140-8985-d263769da28b, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 654.078404] env[63293]: DEBUG nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 654.078537] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "refresh_cache-510b9ea6-04f1-4704-b5e4-e19db57769a8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.078932] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquired lock "refresh_cache-510b9ea6-04f1-4704-b5e4-e19db57769a8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.078973] env[63293]: DEBUG nova.network.neutron [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.313300] env[63293]: DEBUG nova.network.neutron [-] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.606416] env[63293]: DEBUG nova.network.neutron [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.638286] env[63293]: DEBUG nova.compute.manager [req-3612a96c-0931-4ae6-9c50-663ca991e690 req-c33db0c2-135d-4bae-b2aa-ef154e75fe6b service nova] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Received event network-changed-6683eea7-366f-4165-846e-f1103c0a0264 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.638286] env[63293]: DEBUG nova.compute.manager [req-3612a96c-0931-4ae6-9c50-663ca991e690 req-c33db0c2-135d-4bae-b2aa-ef154e75fe6b service nova] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Refreshing instance network info cache due to event network-changed-6683eea7-366f-4165-846e-f1103c0a0264. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 654.638286] env[63293]: DEBUG oslo_concurrency.lockutils [req-3612a96c-0931-4ae6-9c50-663ca991e690 req-c33db0c2-135d-4bae-b2aa-ef154e75fe6b service nova] Acquiring lock "refresh_cache-f5c47736-381f-4979-b804-494ca8cf16a1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.638286] env[63293]: DEBUG oslo_concurrency.lockutils [req-3612a96c-0931-4ae6-9c50-663ca991e690 req-c33db0c2-135d-4bae-b2aa-ef154e75fe6b service nova] Acquired lock "refresh_cache-f5c47736-381f-4979-b804-494ca8cf16a1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.638286] env[63293]: DEBUG nova.network.neutron [req-3612a96c-0931-4ae6-9c50-663ca991e690 req-c33db0c2-135d-4bae-b2aa-ef154e75fe6b service nova] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Refreshing network info cache for port 6683eea7-366f-4165-846e-f1103c0a0264 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 654.777103] env[63293]: DEBUG nova.network.neutron [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.816347] env[63293]: INFO nova.compute.manager [-] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Took 1.04 seconds to deallocate network for instance. [ 654.903989] env[63293]: ERROR nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6683eea7-366f-4165-846e-f1103c0a0264, please check neutron logs for more information. [ 654.903989] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 654.903989] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.903989] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 654.903989] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.903989] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 654.903989] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.903989] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 654.903989] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.903989] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 654.903989] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.903989] env[63293]: ERROR nova.compute.manager raise self.value [ 654.903989] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.903989] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 654.903989] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.903989] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 654.904382] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.904382] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 654.904382] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6683eea7-366f-4165-846e-f1103c0a0264, please check neutron logs for more information. [ 654.904382] env[63293]: ERROR nova.compute.manager [ 654.904382] env[63293]: Traceback (most recent call last): [ 654.904382] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 654.904382] env[63293]: listener.cb(fileno) [ 654.904382] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.904382] env[63293]: result = function(*args, **kwargs) [ 654.904382] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.904382] env[63293]: return func(*args, **kwargs) [ 654.904382] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.904382] env[63293]: raise e [ 654.904382] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.904382] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 654.904382] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.904382] env[63293]: created_port_ids = self._update_ports_for_instance( [ 654.904382] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.904382] env[63293]: with excutils.save_and_reraise_exception(): [ 654.904382] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.904382] env[63293]: self.force_reraise() [ 654.904382] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.904382] env[63293]: raise self.value [ 654.904382] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.904382] env[63293]: updated_port = self._update_port( [ 654.904382] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.904382] env[63293]: _ensure_no_port_binding_failure(port) [ 654.904382] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.904382] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 654.905039] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 6683eea7-366f-4165-846e-f1103c0a0264, please check neutron logs for more information. [ 654.905039] env[63293]: Removing descriptor: 15 [ 654.905320] env[63293]: ERROR nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6683eea7-366f-4165-846e-f1103c0a0264, please check neutron logs for more information. [ 654.905320] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Traceback (most recent call last): [ 654.905320] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 654.905320] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] yield resources [ 654.905320] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.905320] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] self.driver.spawn(context, instance, image_meta, [ 654.905320] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 654.905320] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.905320] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.905320] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] vm_ref = self.build_virtual_machine(instance, [ 654.905320] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] for vif in network_info: [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] return self._sync_wrapper(fn, *args, **kwargs) [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] self.wait() [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] self[:] = self._gt.wait() [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] return self._exit_event.wait() [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.905597] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] result = hub.switch() [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] return self.greenlet.switch() [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] result = function(*args, **kwargs) [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] return func(*args, **kwargs) [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] raise e [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] nwinfo = self.network_api.allocate_for_instance( [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] created_port_ids = self._update_ports_for_instance( [ 654.905906] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] with excutils.save_and_reraise_exception(): [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] self.force_reraise() [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] raise self.value [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] updated_port = self._update_port( [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] _ensure_no_port_binding_failure(port) [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] raise exception.PortBindingFailed(port_id=port['id']) [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] nova.exception.PortBindingFailed: Binding failed for port 6683eea7-366f-4165-846e-f1103c0a0264, please check neutron logs for more information. [ 654.906246] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] [ 654.906618] env[63293]: INFO nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Terminating instance [ 654.907713] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-f5c47736-381f-4979-b804-494ca8cf16a1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.163447] env[63293]: DEBUG nova.network.neutron [req-3612a96c-0931-4ae6-9c50-663ca991e690 req-c33db0c2-135d-4bae-b2aa-ef154e75fe6b service nova] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.262159] env[63293]: DEBUG nova.network.neutron [req-3612a96c-0931-4ae6-9c50-663ca991e690 req-c33db0c2-135d-4bae-b2aa-ef154e75fe6b service nova] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.279888] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Releasing lock "refresh_cache-510b9ea6-04f1-4704-b5e4-e19db57769a8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.280123] env[63293]: DEBUG nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 655.280305] env[63293]: DEBUG nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.280469] env[63293]: DEBUG nova.network.neutron [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.305072] env[63293]: DEBUG nova.network.neutron [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.322823] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.483154] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1480b819-d5e5-4778-85d7-66c41df242d7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.490964] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79de4ed-7180-426d-aaad-b9eac9fe1f35 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.523281] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a9752d-6f8e-431d-b24e-f7605248b4bc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.530763] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c4cafb-8319-483a-b5e1-b53d3e4d047a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.549893] env[63293]: DEBUG nova.compute.provider_tree [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.764791] env[63293]: DEBUG oslo_concurrency.lockutils [req-3612a96c-0931-4ae6-9c50-663ca991e690 req-c33db0c2-135d-4bae-b2aa-ef154e75fe6b service nova] Releasing lock "refresh_cache-f5c47736-381f-4979-b804-494ca8cf16a1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.765240] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-f5c47736-381f-4979-b804-494ca8cf16a1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.765421] env[63293]: DEBUG nova.network.neutron [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 655.807866] env[63293]: DEBUG nova.network.neutron [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.054380] env[63293]: DEBUG nova.scheduler.client.report [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.287972] env[63293]: DEBUG nova.network.neutron [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.313538] env[63293]: INFO nova.compute.manager [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: 510b9ea6-04f1-4704-b5e4-e19db57769a8] Took 1.03 seconds to deallocate network for instance. [ 656.390202] env[63293]: DEBUG nova.network.neutron [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.562921] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.563272] env[63293]: DEBUG nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 656.566335] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.965s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.790557] env[63293]: DEBUG nova.compute.manager [req-c626a598-d43b-48de-b445-15b87211fec3 req-b4641527-25c1-4358-b7b7-3484f9462533 service nova] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Received event network-vif-deleted-6683eea7-366f-4165-846e-f1103c0a0264 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.897041] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-f5c47736-381f-4979-b804-494ca8cf16a1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.897717] env[63293]: DEBUG nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 656.897935] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 656.898267] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4a9543d8-e24a-4006-b031-cf22f4ccc031 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.908212] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15be3773-3993-4128-997b-ce7e74044328 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.934828] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f5c47736-381f-4979-b804-494ca8cf16a1 could not be found. [ 656.934828] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 656.934828] env[63293]: INFO nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 656.934828] env[63293]: DEBUG oslo.service.loopingcall [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 656.934828] env[63293]: DEBUG nova.compute.manager [-] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.934828] env[63293]: DEBUG nova.network.neutron [-] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 656.951119] env[63293]: DEBUG nova.network.neutron [-] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.075025] env[63293]: DEBUG nova.compute.utils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 657.078487] env[63293]: DEBUG nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 657.079307] env[63293]: DEBUG nova.network.neutron [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 657.124664] env[63293]: DEBUG nova.policy [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6860a9e34a1b44029cf57d9f04ae87c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fac34e49d5dc49e7a7055c998d5b6766', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 657.348025] env[63293]: INFO nova.scheduler.client.report [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Deleted allocations for instance 510b9ea6-04f1-4704-b5e4-e19db57769a8 [ 657.454864] env[63293]: DEBUG nova.network.neutron [-] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.509842] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Acquiring lock "52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.510181] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Lock "52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.516284] env[63293]: DEBUG nova.network.neutron [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Successfully created port: 651b6bbf-b272-429d-8a2f-a0f7eae93e81 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 657.567874] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc61b5c-2796-4263-880a-4b5ec084e51a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.576178] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d135c2bc-9f07-4f52-9516-ae48f97de337 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.581524] env[63293]: DEBUG nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 657.609709] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cfa1e65-03c3-45d6-b088-8dbedf9b0dd8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.617800] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-838ab0ef-ad37-4cdb-8d59-2873377fbc6e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.633789] env[63293]: DEBUG nova.compute.provider_tree [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.863087] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27ba0b-737d-47cf-bd7d-156110d54125 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "510b9ea6-04f1-4704-b5e4-e19db57769a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.566s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.960958] env[63293]: INFO nova.compute.manager [-] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Took 1.03 seconds to deallocate network for instance. [ 657.963656] env[63293]: DEBUG nova.compute.claims [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 657.963656] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.140221] env[63293]: DEBUG nova.scheduler.client.report [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.363181] env[63293]: DEBUG nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 658.616113] env[63293]: DEBUG nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 658.646122] env[63293]: DEBUG nova.virt.hardware [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 658.646256] env[63293]: DEBUG nova.virt.hardware [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 658.646671] env[63293]: DEBUG nova.virt.hardware [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 658.646876] env[63293]: DEBUG nova.virt.hardware [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 658.647127] env[63293]: DEBUG nova.virt.hardware [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 658.647320] env[63293]: DEBUG nova.virt.hardware [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 658.647580] env[63293]: DEBUG nova.virt.hardware [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 658.647798] env[63293]: DEBUG nova.virt.hardware [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 658.648023] env[63293]: DEBUG nova.virt.hardware [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 658.648197] env[63293]: DEBUG nova.virt.hardware [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 658.648401] env[63293]: DEBUG nova.virt.hardware [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 658.649282] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.083s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.649971] env[63293]: ERROR nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 453608e9-ae55-48dd-b58b-6189194d69c0, please check neutron logs for more information. [ 658.649971] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Traceback (most recent call last): [ 658.649971] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 658.649971] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] self.driver.spawn(context, instance, image_meta, [ 658.649971] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 658.649971] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.649971] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.649971] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] vm_ref = self.build_virtual_machine(instance, [ 658.649971] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.649971] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.649971] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] for vif in network_info: [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] return self._sync_wrapper(fn, *args, **kwargs) [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] self.wait() [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] self[:] = self._gt.wait() [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] return self._exit_event.wait() [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] result = hub.switch() [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 658.650306] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] return self.greenlet.switch() [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] result = function(*args, **kwargs) [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] return func(*args, **kwargs) [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] raise e [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] nwinfo = self.network_api.allocate_for_instance( [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] created_port_ids = self._update_ports_for_instance( [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] with excutils.save_and_reraise_exception(): [ 658.650641] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.650968] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] self.force_reraise() [ 658.650968] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.650968] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] raise self.value [ 658.650968] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.650968] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] updated_port = self._update_port( [ 658.650968] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.650968] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] _ensure_no_port_binding_failure(port) [ 658.650968] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.650968] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] raise exception.PortBindingFailed(port_id=port['id']) [ 658.650968] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] nova.exception.PortBindingFailed: Binding failed for port 453608e9-ae55-48dd-b58b-6189194d69c0, please check neutron logs for more information. [ 658.650968] env[63293]: ERROR nova.compute.manager [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] [ 658.651409] env[63293]: DEBUG nova.compute.utils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Binding failed for port 453608e9-ae55-48dd-b58b-6189194d69c0, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 658.653726] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0900be0b-86c6-4346-b407-22e76f1e9667 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.657480] env[63293]: DEBUG nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Build of instance eaaf2515-cc4b-42c7-aacd-14074c760924 was re-scheduled: Binding failed for port 453608e9-ae55-48dd-b58b-6189194d69c0, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 658.657991] env[63293]: DEBUG nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 658.658274] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "refresh_cache-eaaf2515-cc4b-42c7-aacd-14074c760924" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.658463] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquired lock "refresh_cache-eaaf2515-cc4b-42c7-aacd-14074c760924" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.658813] env[63293]: DEBUG nova.network.neutron [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.663636] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.291s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.674237] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6840696c-23f0-4372-b22e-8907edbb5b43 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.840548] env[63293]: ERROR nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 651b6bbf-b272-429d-8a2f-a0f7eae93e81, please check neutron logs for more information. [ 658.840548] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 658.840548] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.840548] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 658.840548] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.840548] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 658.840548] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.840548] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 658.840548] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.840548] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 658.840548] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.840548] env[63293]: ERROR nova.compute.manager raise self.value [ 658.840548] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.840548] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 658.840548] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.840548] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 658.841024] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.841024] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 658.841024] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 651b6bbf-b272-429d-8a2f-a0f7eae93e81, please check neutron logs for more information. [ 658.841024] env[63293]: ERROR nova.compute.manager [ 658.841024] env[63293]: Traceback (most recent call last): [ 658.841024] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 658.841024] env[63293]: listener.cb(fileno) [ 658.841024] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.841024] env[63293]: result = function(*args, **kwargs) [ 658.841024] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 658.841024] env[63293]: return func(*args, **kwargs) [ 658.841024] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.841024] env[63293]: raise e [ 658.841024] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.841024] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 658.841024] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.841024] env[63293]: created_port_ids = self._update_ports_for_instance( [ 658.841024] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.841024] env[63293]: with excutils.save_and_reraise_exception(): [ 658.841024] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.841024] env[63293]: self.force_reraise() [ 658.841024] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.841024] env[63293]: raise self.value [ 658.841024] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.841024] env[63293]: updated_port = self._update_port( [ 658.841024] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.841024] env[63293]: _ensure_no_port_binding_failure(port) [ 658.841024] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.841024] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 658.841815] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 651b6bbf-b272-429d-8a2f-a0f7eae93e81, please check neutron logs for more information. [ 658.841815] env[63293]: Removing descriptor: 15 [ 658.841815] env[63293]: ERROR nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 651b6bbf-b272-429d-8a2f-a0f7eae93e81, please check neutron logs for more information. [ 658.841815] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Traceback (most recent call last): [ 658.841815] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 658.841815] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] yield resources [ 658.841815] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 658.841815] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] self.driver.spawn(context, instance, image_meta, [ 658.841815] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 658.841815] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.841815] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.841815] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] vm_ref = self.build_virtual_machine(instance, [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] for vif in network_info: [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] return self._sync_wrapper(fn, *args, **kwargs) [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] self.wait() [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] self[:] = self._gt.wait() [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] return self._exit_event.wait() [ 658.842177] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] result = hub.switch() [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] return self.greenlet.switch() [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] result = function(*args, **kwargs) [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] return func(*args, **kwargs) [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] raise e [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] nwinfo = self.network_api.allocate_for_instance( [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.842593] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] created_port_ids = self._update_ports_for_instance( [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] with excutils.save_and_reraise_exception(): [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] self.force_reraise() [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] raise self.value [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] updated_port = self._update_port( [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] _ensure_no_port_binding_failure(port) [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.842953] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] raise exception.PortBindingFailed(port_id=port['id']) [ 658.843393] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] nova.exception.PortBindingFailed: Binding failed for port 651b6bbf-b272-429d-8a2f-a0f7eae93e81, please check neutron logs for more information. [ 658.843393] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] [ 658.843393] env[63293]: INFO nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Terminating instance [ 658.844198] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-947f69fa-77cf-4c7d-ba62-37567318b969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.844383] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-947f69fa-77cf-4c7d-ba62-37567318b969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.844566] env[63293]: DEBUG nova.network.neutron [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.885569] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.939593] env[63293]: DEBUG nova.compute.manager [req-bd99638e-113b-42c8-bb2a-db5ad6168b0b req-648a5083-9a7b-4183-8723-bc1ee24a3084 service nova] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Received event network-changed-651b6bbf-b272-429d-8a2f-a0f7eae93e81 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 658.939895] env[63293]: DEBUG nova.compute.manager [req-bd99638e-113b-42c8-bb2a-db5ad6168b0b req-648a5083-9a7b-4183-8723-bc1ee24a3084 service nova] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Refreshing instance network info cache due to event network-changed-651b6bbf-b272-429d-8a2f-a0f7eae93e81. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 658.940112] env[63293]: DEBUG oslo_concurrency.lockutils [req-bd99638e-113b-42c8-bb2a-db5ad6168b0b req-648a5083-9a7b-4183-8723-bc1ee24a3084 service nova] Acquiring lock "refresh_cache-947f69fa-77cf-4c7d-ba62-37567318b969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.179773] env[63293]: DEBUG nova.network.neutron [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.277747] env[63293]: DEBUG nova.network.neutron [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.365069] env[63293]: DEBUG nova.network.neutron [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.386738] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "b48324a4-902d-4642-8827-daee9683b3a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.386944] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "b48324a4-902d-4642-8827-daee9683b3a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.445327] env[63293]: DEBUG nova.network.neutron [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.560621] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23dd05dd-2c60-487f-8024-02e92b8261c7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.568277] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015ef194-c50f-4399-a227-1e6480c87648 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.598027] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc42817-23dd-489d-91f7-31673298bd47 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.604967] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa07bd3-f804-4fee-9acc-ddff8159e026 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.617764] env[63293]: DEBUG nova.compute.provider_tree [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.780661] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Releasing lock "refresh_cache-eaaf2515-cc4b-42c7-aacd-14074c760924" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.780917] env[63293]: DEBUG nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 659.781111] env[63293]: DEBUG nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.781277] env[63293]: DEBUG nova.network.neutron [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 659.796826] env[63293]: DEBUG nova.network.neutron [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.949323] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-947f69fa-77cf-4c7d-ba62-37567318b969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.949772] env[63293]: DEBUG nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 659.950049] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 659.950359] env[63293]: DEBUG oslo_concurrency.lockutils [req-bd99638e-113b-42c8-bb2a-db5ad6168b0b req-648a5083-9a7b-4183-8723-bc1ee24a3084 service nova] Acquired lock "refresh_cache-947f69fa-77cf-4c7d-ba62-37567318b969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.950533] env[63293]: DEBUG nova.network.neutron [req-bd99638e-113b-42c8-bb2a-db5ad6168b0b req-648a5083-9a7b-4183-8723-bc1ee24a3084 service nova] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Refreshing network info cache for port 651b6bbf-b272-429d-8a2f-a0f7eae93e81 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 659.951933] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86225ea4-40a9-48d1-a61f-0ea72ebfbe46 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.961584] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7189692e-b402-4afe-8908-45d10c8402d6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.983274] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 947f69fa-77cf-4c7d-ba62-37567318b969 could not be found. [ 659.983497] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 659.983674] env[63293]: INFO nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Took 0.03 seconds to destroy the instance on the hypervisor. [ 659.983912] env[63293]: DEBUG oslo.service.loopingcall [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 659.984145] env[63293]: DEBUG nova.compute.manager [-] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.984239] env[63293]: DEBUG nova.network.neutron [-] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 659.997716] env[63293]: DEBUG nova.network.neutron [-] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.120383] env[63293]: DEBUG nova.scheduler.client.report [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.300041] env[63293]: DEBUG nova.network.neutron [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.472029] env[63293]: DEBUG nova.network.neutron [req-bd99638e-113b-42c8-bb2a-db5ad6168b0b req-648a5083-9a7b-4183-8723-bc1ee24a3084 service nova] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.500588] env[63293]: DEBUG nova.network.neutron [-] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.551630] env[63293]: DEBUG nova.network.neutron [req-bd99638e-113b-42c8-bb2a-db5ad6168b0b req-648a5083-9a7b-4183-8723-bc1ee24a3084 service nova] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.626621] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.627289] env[63293]: ERROR nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 255fd7f1-f9a0-4469-80e0-980345a3cb6d, please check neutron logs for more information. [ 660.627289] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Traceback (most recent call last): [ 660.627289] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.627289] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] self.driver.spawn(context, instance, image_meta, [ 660.627289] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 660.627289] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.627289] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.627289] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] vm_ref = self.build_virtual_machine(instance, [ 660.627289] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.627289] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.627289] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] for vif in network_info: [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] return self._sync_wrapper(fn, *args, **kwargs) [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] self.wait() [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] self[:] = self._gt.wait() [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] return self._exit_event.wait() [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] result = hub.switch() [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.627606] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] return self.greenlet.switch() [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] result = function(*args, **kwargs) [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] return func(*args, **kwargs) [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] raise e [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] nwinfo = self.network_api.allocate_for_instance( [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] created_port_ids = self._update_ports_for_instance( [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] with excutils.save_and_reraise_exception(): [ 660.627915] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.628251] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] self.force_reraise() [ 660.628251] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.628251] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] raise self.value [ 660.628251] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.628251] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] updated_port = self._update_port( [ 660.628251] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.628251] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] _ensure_no_port_binding_failure(port) [ 660.628251] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.628251] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] raise exception.PortBindingFailed(port_id=port['id']) [ 660.628251] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] nova.exception.PortBindingFailed: Binding failed for port 255fd7f1-f9a0-4469-80e0-980345a3cb6d, please check neutron logs for more information. [ 660.628251] env[63293]: ERROR nova.compute.manager [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] [ 660.628578] env[63293]: DEBUG nova.compute.utils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Binding failed for port 255fd7f1-f9a0-4469-80e0-980345a3cb6d, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 660.629192] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.764s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.632838] env[63293]: DEBUG nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Build of instance 1814ebc5-7869-4ec6-8cf6-790daed5e271 was re-scheduled: Binding failed for port 255fd7f1-f9a0-4469-80e0-980345a3cb6d, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 660.633028] env[63293]: DEBUG nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 660.633371] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "refresh_cache-1814ebc5-7869-4ec6-8cf6-790daed5e271" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.633517] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "refresh_cache-1814ebc5-7869-4ec6-8cf6-790daed5e271" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.633676] env[63293]: DEBUG nova.network.neutron [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 660.802622] env[63293]: INFO nova.compute.manager [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: eaaf2515-cc4b-42c7-aacd-14074c760924] Took 1.02 seconds to deallocate network for instance. [ 660.991708] env[63293]: DEBUG nova.compute.manager [req-de0e81e4-091c-4402-85d5-abef5d110e87 req-babc8fe7-f4da-48b1-8d27-056b40e520e5 service nova] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Received event network-vif-deleted-651b6bbf-b272-429d-8a2f-a0f7eae93e81 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 661.004062] env[63293]: INFO nova.compute.manager [-] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Took 1.02 seconds to deallocate network for instance. [ 661.006212] env[63293]: DEBUG nova.compute.claims [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 661.006390] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.054161] env[63293]: DEBUG oslo_concurrency.lockutils [req-bd99638e-113b-42c8-bb2a-db5ad6168b0b req-648a5083-9a7b-4183-8723-bc1ee24a3084 service nova] Releasing lock "refresh_cache-947f69fa-77cf-4c7d-ba62-37567318b969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.157828] env[63293]: DEBUG nova.network.neutron [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.243202] env[63293]: DEBUG nova.network.neutron [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.480071] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448d7ff0-fb2b-4be7-8342-517437e720c2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.488276] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba048aa-2f15-4169-bbc3-f287bf933703 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.518988] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec843c7-87b3-425e-b4a2-6cbde8c231ad {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.527963] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a35bd4-aad0-4459-a37d-13cef9824a1e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.542423] env[63293]: DEBUG nova.compute.provider_tree [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.746036] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "refresh_cache-1814ebc5-7869-4ec6-8cf6-790daed5e271" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.746300] env[63293]: DEBUG nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 661.746481] env[63293]: DEBUG nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.746649] env[63293]: DEBUG nova.network.neutron [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.765823] env[63293]: DEBUG nova.network.neutron [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.830594] env[63293]: INFO nova.scheduler.client.report [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Deleted allocations for instance eaaf2515-cc4b-42c7-aacd-14074c760924 [ 662.046033] env[63293]: DEBUG nova.scheduler.client.report [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.268891] env[63293]: DEBUG nova.network.neutron [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.339152] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4299d36d-ffd4-4dfb-a2c4-c871e0c11766 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "eaaf2515-cc4b-42c7-aacd-14074c760924" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.905s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.551247] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.551896] env[63293]: ERROR nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 95e989b9-668a-43ee-8cdd-9069454e4e1c, please check neutron logs for more information. [ 662.551896] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Traceback (most recent call last): [ 662.551896] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 662.551896] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] self.driver.spawn(context, instance, image_meta, [ 662.551896] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 662.551896] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.551896] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.551896] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] vm_ref = self.build_virtual_machine(instance, [ 662.551896] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.551896] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.551896] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] for vif in network_info: [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] return self._sync_wrapper(fn, *args, **kwargs) [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] self.wait() [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] self[:] = self._gt.wait() [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] return self._exit_event.wait() [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] result = hub.switch() [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 662.552210] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] return self.greenlet.switch() [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] result = function(*args, **kwargs) [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] return func(*args, **kwargs) [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] raise e [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] nwinfo = self.network_api.allocate_for_instance( [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] created_port_ids = self._update_ports_for_instance( [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] with excutils.save_and_reraise_exception(): [ 662.552511] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.552817] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] self.force_reraise() [ 662.552817] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.552817] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] raise self.value [ 662.552817] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.552817] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] updated_port = self._update_port( [ 662.552817] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.552817] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] _ensure_no_port_binding_failure(port) [ 662.552817] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.552817] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] raise exception.PortBindingFailed(port_id=port['id']) [ 662.552817] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] nova.exception.PortBindingFailed: Binding failed for port 95e989b9-668a-43ee-8cdd-9069454e4e1c, please check neutron logs for more information. [ 662.552817] env[63293]: ERROR nova.compute.manager [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] [ 662.553145] env[63293]: DEBUG nova.compute.utils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Binding failed for port 95e989b9-668a-43ee-8cdd-9069454e4e1c, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 662.553823] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.201s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.555358] env[63293]: INFO nova.compute.claims [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 662.558241] env[63293]: DEBUG nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Build of instance 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5 was re-scheduled: Binding failed for port 95e989b9-668a-43ee-8cdd-9069454e4e1c, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 662.558698] env[63293]: DEBUG nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 662.558931] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Acquiring lock "refresh_cache-63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.559092] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Acquired lock "refresh_cache-63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.559251] env[63293]: DEBUG nova.network.neutron [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 662.772044] env[63293]: INFO nova.compute.manager [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 1814ebc5-7869-4ec6-8cf6-790daed5e271] Took 1.03 seconds to deallocate network for instance. [ 662.840962] env[63293]: DEBUG nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 663.078356] env[63293]: DEBUG nova.network.neutron [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.189537] env[63293]: DEBUG nova.network.neutron [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.386340] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.693909] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Releasing lock "refresh_cache-63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.694170] env[63293]: DEBUG nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 663.694352] env[63293]: DEBUG nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.694515] env[63293]: DEBUG nova.network.neutron [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 663.711676] env[63293]: DEBUG nova.network.neutron [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.803767] env[63293]: INFO nova.scheduler.client.report [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleted allocations for instance 1814ebc5-7869-4ec6-8cf6-790daed5e271 [ 663.929138] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0b876c-7797-49c5-b605-ca06d38538e5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.938706] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a9b99d-3f75-45e2-b3a4-2de6eec3ee9c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.968837] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd5ff49-e04b-43c7-8719-27a57111a830 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.976167] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dba1a20-393e-4f5c-a5da-006dea2965b3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.989168] env[63293]: DEBUG nova.compute.provider_tree [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.214683] env[63293]: DEBUG nova.network.neutron [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.313958] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a3b644d9-41c8-4d0c-bf15-9441b3f18dbf tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "1814ebc5-7869-4ec6-8cf6-790daed5e271" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.085s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.492334] env[63293]: DEBUG nova.scheduler.client.report [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.717802] env[63293]: INFO nova.compute.manager [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] [instance: 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5] Took 1.02 seconds to deallocate network for instance. [ 664.817052] env[63293]: DEBUG nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 664.883281] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "495cb1b0-9194-44e6-9f68-de62033ea17d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.883504] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "495cb1b0-9194-44e6-9f68-de62033ea17d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.998291] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.999469] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.382s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.335927] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.503907] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Acquiring lock "f0cddf65-e5d8-434d-88b8-f68f044ba91f" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.504269] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Lock "f0cddf65-e5d8-434d-88b8-f68f044ba91f" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.742028] env[63293]: INFO nova.scheduler.client.report [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Deleted allocations for instance 63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5 [ 665.872896] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.873139] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.873882] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fb91f89-3ea1-4317-86f6-d63a272b0d4d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.882069] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29288a26-11f9-4684-98a9-96232fd3dcdd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.912633] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6d5517-b736-439a-b5bc-7e7efffe7b7c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.919545] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e57fb25-3f89-4477-8735-3e0f74e4e7b6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.932874] env[63293]: DEBUG nova.compute.provider_tree [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.010941] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Lock "f0cddf65-e5d8-434d-88b8-f68f044ba91f" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.507s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.011632] env[63293]: DEBUG nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.255194] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2fdda777-6bfa-42a3-acf1-95ae1962716d tempest-ServerActionsTestOtherA-1746632528 tempest-ServerActionsTestOtherA-1746632528-project-member] Lock "63f13b3b-d0e3-46d7-ab73-c6bf04d6f1d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.110s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.436721] env[63293]: DEBUG nova.scheduler.client.report [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.517108] env[63293]: DEBUG nova.compute.utils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 666.518967] env[63293]: DEBUG nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 666.521267] env[63293]: DEBUG nova.network.neutron [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 666.580967] env[63293]: DEBUG nova.policy [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d25fcbdacae4467b2ebf37f5f6d99ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a260f2cc36004933a65cd2d1811b930d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 666.758215] env[63293]: DEBUG nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 666.942298] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.942972] env[63293]: ERROR nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9f72114f-ae1c-4928-8d06-bc85d5cc0a96, please check neutron logs for more information. [ 666.942972] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Traceback (most recent call last): [ 666.942972] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.942972] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] self.driver.spawn(context, instance, image_meta, [ 666.942972] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 666.942972] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.942972] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.942972] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] vm_ref = self.build_virtual_machine(instance, [ 666.942972] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.942972] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.942972] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] for vif in network_info: [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] return self._sync_wrapper(fn, *args, **kwargs) [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] self.wait() [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] self[:] = self._gt.wait() [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] return self._exit_event.wait() [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] result = hub.switch() [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.943311] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] return self.greenlet.switch() [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] result = function(*args, **kwargs) [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] return func(*args, **kwargs) [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] raise e [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] nwinfo = self.network_api.allocate_for_instance( [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] created_port_ids = self._update_ports_for_instance( [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] with excutils.save_and_reraise_exception(): [ 666.943637] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.944083] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] self.force_reraise() [ 666.944083] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.944083] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] raise self.value [ 666.944083] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.944083] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] updated_port = self._update_port( [ 666.944083] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.944083] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] _ensure_no_port_binding_failure(port) [ 666.944083] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.944083] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] raise exception.PortBindingFailed(port_id=port['id']) [ 666.944083] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] nova.exception.PortBindingFailed: Binding failed for port 9f72114f-ae1c-4928-8d06-bc85d5cc0a96, please check neutron logs for more information. [ 666.944083] env[63293]: ERROR nova.compute.manager [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] [ 666.944350] env[63293]: DEBUG nova.compute.utils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Binding failed for port 9f72114f-ae1c-4928-8d06-bc85d5cc0a96, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 666.945528] env[63293]: DEBUG nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Build of instance 87c7bb2c-46ff-45f7-8967-14192f2b42e4 was re-scheduled: Binding failed for port 9f72114f-ae1c-4928-8d06-bc85d5cc0a96, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 666.945978] env[63293]: DEBUG nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 666.946232] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquiring lock "refresh_cache-87c7bb2c-46ff-45f7-8967-14192f2b42e4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.946382] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquired lock "refresh_cache-87c7bb2c-46ff-45f7-8967-14192f2b42e4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.946540] env[63293]: DEBUG nova.network.neutron [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 666.948202] env[63293]: DEBUG nova.network.neutron [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Successfully created port: ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 666.952184] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.851s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.024568] env[63293]: DEBUG nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 667.288570] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.475326] env[63293]: DEBUG nova.network.neutron [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.535133] env[63293]: DEBUG nova.network.neutron [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.899272] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f279735-b223-4418-bc5d-48936d04caf1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.908026] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b74877-7670-491e-a5a8-e0bddb916216 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.940270] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e32b9cf-96de-47dc-98d9-ddced19fa736 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.953018] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378f8eba-507d-415a-ab8c-5b5bc24d976a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.963907] env[63293]: DEBUG nova.compute.provider_tree [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.041253] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Releasing lock "refresh_cache-87c7bb2c-46ff-45f7-8967-14192f2b42e4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.041253] env[63293]: DEBUG nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 668.041253] env[63293]: DEBUG nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 668.041253] env[63293]: DEBUG nova.network.neutron [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 668.041253] env[63293]: DEBUG nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.072145] env[63293]: DEBUG nova.network.neutron [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.075942] env[63293]: DEBUG nova.virt.hardware [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.076351] env[63293]: DEBUG nova.virt.hardware [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.079018] env[63293]: DEBUG nova.virt.hardware [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.079018] env[63293]: DEBUG nova.virt.hardware [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.079018] env[63293]: DEBUG nova.virt.hardware [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.079018] env[63293]: DEBUG nova.virt.hardware [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.079018] env[63293]: DEBUG nova.virt.hardware [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.079267] env[63293]: DEBUG nova.virt.hardware [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.079305] env[63293]: DEBUG nova.virt.hardware [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.079512] env[63293]: DEBUG nova.virt.hardware [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.079807] env[63293]: DEBUG nova.virt.hardware [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.080742] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ac264f-736c-471d-b049-b3315bf61058 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.084546] env[63293]: DEBUG nova.network.neutron [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.092560] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ba8267-9115-4ddf-ba45-e93439a6a87d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.468032] env[63293]: DEBUG nova.scheduler.client.report [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.588252] env[63293]: INFO nova.compute.manager [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: 87c7bb2c-46ff-45f7-8967-14192f2b42e4] Took 0.55 seconds to deallocate network for instance. [ 668.613558] env[63293]: DEBUG nova.compute.manager [req-539f0693-4cdd-4592-b796-a6d94ccba5dc req-492edf65-3078-4693-bc81-5ec7975dd423 service nova] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Received event network-changed-ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 668.613558] env[63293]: DEBUG nova.compute.manager [req-539f0693-4cdd-4592-b796-a6d94ccba5dc req-492edf65-3078-4693-bc81-5ec7975dd423 service nova] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Refreshing instance network info cache due to event network-changed-ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 668.613558] env[63293]: DEBUG oslo_concurrency.lockutils [req-539f0693-4cdd-4592-b796-a6d94ccba5dc req-492edf65-3078-4693-bc81-5ec7975dd423 service nova] Acquiring lock "refresh_cache-dc9ae29a-a61b-4402-a16f-460cf74bc10c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.613558] env[63293]: DEBUG oslo_concurrency.lockutils [req-539f0693-4cdd-4592-b796-a6d94ccba5dc req-492edf65-3078-4693-bc81-5ec7975dd423 service nova] Acquired lock "refresh_cache-dc9ae29a-a61b-4402-a16f-460cf74bc10c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.613558] env[63293]: DEBUG nova.network.neutron [req-539f0693-4cdd-4592-b796-a6d94ccba5dc req-492edf65-3078-4693-bc81-5ec7975dd423 service nova] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Refreshing network info cache for port ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 668.874534] env[63293]: ERROR nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0, please check neutron logs for more information. [ 668.874534] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 668.874534] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.874534] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 668.874534] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.874534] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 668.874534] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.874534] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 668.874534] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.874534] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 668.874534] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.874534] env[63293]: ERROR nova.compute.manager raise self.value [ 668.874534] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.874534] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 668.874534] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.874534] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 668.874977] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.874977] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 668.874977] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0, please check neutron logs for more information. [ 668.874977] env[63293]: ERROR nova.compute.manager [ 668.874977] env[63293]: Traceback (most recent call last): [ 668.874977] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 668.874977] env[63293]: listener.cb(fileno) [ 668.874977] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.874977] env[63293]: result = function(*args, **kwargs) [ 668.874977] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.874977] env[63293]: return func(*args, **kwargs) [ 668.874977] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.874977] env[63293]: raise e [ 668.874977] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.874977] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 668.874977] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.874977] env[63293]: created_port_ids = self._update_ports_for_instance( [ 668.874977] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.874977] env[63293]: with excutils.save_and_reraise_exception(): [ 668.874977] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.874977] env[63293]: self.force_reraise() [ 668.874977] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.874977] env[63293]: raise self.value [ 668.874977] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.874977] env[63293]: updated_port = self._update_port( [ 668.874977] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.874977] env[63293]: _ensure_no_port_binding_failure(port) [ 668.874977] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.874977] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 668.875774] env[63293]: nova.exception.PortBindingFailed: Binding failed for port ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0, please check neutron logs for more information. [ 668.875774] env[63293]: Removing descriptor: 15 [ 668.875835] env[63293]: ERROR nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0, please check neutron logs for more information. [ 668.875835] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Traceback (most recent call last): [ 668.875835] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 668.875835] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] yield resources [ 668.875835] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.875835] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] self.driver.spawn(context, instance, image_meta, [ 668.875835] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 668.875835] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.875835] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.875835] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] vm_ref = self.build_virtual_machine(instance, [ 668.875835] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] for vif in network_info: [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] return self._sync_wrapper(fn, *args, **kwargs) [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] self.wait() [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] self[:] = self._gt.wait() [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] return self._exit_event.wait() [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 668.876123] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] result = hub.switch() [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] return self.greenlet.switch() [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] result = function(*args, **kwargs) [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] return func(*args, **kwargs) [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] raise e [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] nwinfo = self.network_api.allocate_for_instance( [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] created_port_ids = self._update_ports_for_instance( [ 668.876426] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] with excutils.save_and_reraise_exception(): [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] self.force_reraise() [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] raise self.value [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] updated_port = self._update_port( [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] _ensure_no_port_binding_failure(port) [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] raise exception.PortBindingFailed(port_id=port['id']) [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] nova.exception.PortBindingFailed: Binding failed for port ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0, please check neutron logs for more information. [ 668.876743] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] [ 668.877075] env[63293]: INFO nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Terminating instance [ 668.881554] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Acquiring lock "refresh_cache-dc9ae29a-a61b-4402-a16f-460cf74bc10c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.974032] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.023s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.975069] env[63293]: ERROR nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 89a49c2d-7349-467b-96ab-5b38bc334f87, please check neutron logs for more information. [ 668.975069] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Traceback (most recent call last): [ 668.975069] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.975069] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] self.driver.spawn(context, instance, image_meta, [ 668.975069] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 668.975069] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.975069] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.975069] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] vm_ref = self.build_virtual_machine(instance, [ 668.975069] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.975069] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.975069] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] for vif in network_info: [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] return self._sync_wrapper(fn, *args, **kwargs) [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] self.wait() [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] self[:] = self._gt.wait() [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] return self._exit_event.wait() [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] result = hub.switch() [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 668.975469] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] return self.greenlet.switch() [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] result = function(*args, **kwargs) [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] return func(*args, **kwargs) [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] raise e [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] nwinfo = self.network_api.allocate_for_instance( [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] created_port_ids = self._update_ports_for_instance( [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] with excutils.save_and_reraise_exception(): [ 668.975794] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.976120] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] self.force_reraise() [ 668.976120] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.976120] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] raise self.value [ 668.976120] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.976120] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] updated_port = self._update_port( [ 668.976120] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.976120] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] _ensure_no_port_binding_failure(port) [ 668.976120] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.976120] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] raise exception.PortBindingFailed(port_id=port['id']) [ 668.976120] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] nova.exception.PortBindingFailed: Binding failed for port 89a49c2d-7349-467b-96ab-5b38bc334f87, please check neutron logs for more information. [ 668.976120] env[63293]: ERROR nova.compute.manager [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] [ 668.976387] env[63293]: DEBUG nova.compute.utils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Binding failed for port 89a49c2d-7349-467b-96ab-5b38bc334f87, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 668.976491] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.929s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.978478] env[63293]: DEBUG nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Build of instance 769c2643-57eb-456c-8083-997a8c9c48cc was re-scheduled: Binding failed for port 89a49c2d-7349-467b-96ab-5b38bc334f87, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 668.978977] env[63293]: DEBUG nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 668.979250] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquiring lock "refresh_cache-769c2643-57eb-456c-8083-997a8c9c48cc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.979436] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Acquired lock "refresh_cache-769c2643-57eb-456c-8083-997a8c9c48cc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.979628] env[63293]: DEBUG nova.network.neutron [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 669.132665] env[63293]: DEBUG nova.network.neutron [req-539f0693-4cdd-4592-b796-a6d94ccba5dc req-492edf65-3078-4693-bc81-5ec7975dd423 service nova] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.264810] env[63293]: DEBUG nova.network.neutron [req-539f0693-4cdd-4592-b796-a6d94ccba5dc req-492edf65-3078-4693-bc81-5ec7975dd423 service nova] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.503032] env[63293]: DEBUG nova.network.neutron [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.574755] env[63293]: DEBUG nova.network.neutron [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.623441] env[63293]: INFO nova.scheduler.client.report [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Deleted allocations for instance 87c7bb2c-46ff-45f7-8967-14192f2b42e4 [ 669.767984] env[63293]: DEBUG oslo_concurrency.lockutils [req-539f0693-4cdd-4592-b796-a6d94ccba5dc req-492edf65-3078-4693-bc81-5ec7975dd423 service nova] Releasing lock "refresh_cache-dc9ae29a-a61b-4402-a16f-460cf74bc10c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.768406] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Acquired lock "refresh_cache-dc9ae29a-a61b-4402-a16f-460cf74bc10c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.768591] env[63293]: DEBUG nova.network.neutron [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 670.007868] env[63293]: WARNING nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance b2055187-edd8-4a8c-9811-183cf7ba3b39 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 670.078276] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Releasing lock "refresh_cache-769c2643-57eb-456c-8083-997a8c9c48cc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.079687] env[63293]: DEBUG nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 670.079687] env[63293]: DEBUG nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.079687] env[63293]: DEBUG nova.network.neutron [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.096840] env[63293]: DEBUG nova.network.neutron [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.133083] env[63293]: DEBUG oslo_concurrency.lockutils [None req-764d1ebd-f57a-46fd-be1b-9f46203b0591 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "87c7bb2c-46ff-45f7-8967-14192f2b42e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.122s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.289049] env[63293]: DEBUG nova.network.neutron [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.361804] env[63293]: DEBUG nova.network.neutron [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.510974] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 769c2643-57eb-456c-8083-997a8c9c48cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.511231] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance b2c1e806-d0ae-455f-95e9-405a9557697f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 670.511422] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance f5c47736-381f-4979-b804-494ca8cf16a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 670.511575] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 947f69fa-77cf-4c7d-ba62-37567318b969 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 670.511753] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance dc9ae29a-a61b-4402-a16f-460cf74bc10c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 670.599764] env[63293]: DEBUG nova.network.neutron [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.637498] env[63293]: DEBUG nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 670.644238] env[63293]: DEBUG nova.compute.manager [req-d9ecd54a-5e17-4e97-ad87-d8458bc636d6 req-52930c88-cf21-4ab9-bc9a-d5d3a8d6de69 service nova] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Received event network-vif-deleted-ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.866062] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Releasing lock "refresh_cache-dc9ae29a-a61b-4402-a16f-460cf74bc10c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.866314] env[63293]: DEBUG nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 670.866507] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 670.866826] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bec06580-9f34-48eb-8c01-fe2118301bb0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.879146] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb1d8df1-2d31-4509-8cbb-f651cbd97250 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.904846] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dc9ae29a-a61b-4402-a16f-460cf74bc10c could not be found. [ 670.905084] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 670.905343] env[63293]: INFO nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 670.905615] env[63293]: DEBUG oslo.service.loopingcall [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 670.905842] env[63293]: DEBUG nova.compute.manager [-] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.905937] env[63293]: DEBUG nova.network.neutron [-] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.921134] env[63293]: DEBUG nova.network.neutron [-] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.017262] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 92208bca-bbac-48fa-83db-3f721e6c80c2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.105960] env[63293]: INFO nova.compute.manager [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] [instance: 769c2643-57eb-456c-8083-997a8c9c48cc] Took 1.03 seconds to deallocate network for instance. [ 671.164162] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.427793] env[63293]: DEBUG nova.network.neutron [-] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.520866] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 305e3a9c-27cf-47b8-8825-09e5abe220ee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.932946] env[63293]: INFO nova.compute.manager [-] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Took 1.03 seconds to deallocate network for instance. [ 671.939015] env[63293]: DEBUG nova.compute.claims [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 671.939015] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.027988] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance a63744c8-b9bf-4a89-9319-061d53de6b5f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.152182] env[63293]: INFO nova.scheduler.client.report [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Deleted allocations for instance 769c2643-57eb-456c-8083-997a8c9c48cc [ 672.532991] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 8b4b041b-5d3f-479a-8241-c75cce650d1e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.662738] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ef852b7-f839-4eff-b7cd-4f44288901f0 tempest-DeleteServersAdminTestJSON-1435620798 tempest-DeleteServersAdminTestJSON-1435620798-project-member] Lock "769c2643-57eb-456c-8083-997a8c9c48cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.604s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.039431] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 8c2439e1-26ce-45e5-9179-164c0dbbbf76 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.167477] env[63293]: DEBUG nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.544975] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance e0e08728-b6c3-4610-bc27-348c5594fefc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.691734] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.812156] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.812399] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.048251] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 2afc50fb-138c-4523-b086-db11026ae52f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.551613] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 5bf150a5-ad49-4bf8-a8fa-e376e0f00926 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.054990] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance e86185ea-7aa6-43ad-920d-645e18194b04 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.557826] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 0aade4da-882a-4342-aefb-07eb306b17ff has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.642638] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquiring lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.642862] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.061859] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance db04bc40-1106-46dd-ad9a-f9b11a6c8774 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 676.565837] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 263d1329-a81a-4d50-91b6-3927bffe8d4c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 677.068907] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance fd67940d-9552-42b3-b472-2ce0e7899ea6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 677.571906] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 678.075847] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 985a5c30-551f-4de7-bc85-d53f002e0f72 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 678.580089] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance b5334832-25dc-4b03-9689-388a1382200a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 679.083054] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 67839a84-c1a0-4728-88f4-47acf99f1fc4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 679.584834] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 0df77939-d60c-4ccb-98f7-2325062ec76f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 680.087772] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance b1b1ebe0-187b-4e23-9433-7dff420e3de3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 680.590698] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.095211] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.599267] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance b48324a4-902d-4642-8827-daee9683b3a0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.102485] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 495cb1b0-9194-44e6-9f68-de62033ea17d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.605553] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.605827] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 682.605988] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 682.928190] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2200b8-5bcd-4f88-a915-4de08d7e0d9a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.936410] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d287231a-db5c-4794-ab51-c4e1a0b3d680 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.967632] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ebe415f-b29b-480e-b797-df833884e1bc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.974797] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ed81ee-6562-4bca-8527-db46e84dc9e3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.988504] env[63293]: DEBUG nova.compute.provider_tree [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.493108] env[63293]: DEBUG nova.scheduler.client.report [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.999986] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 683.999986] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 15.023s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.999986] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.285s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.842703] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8149e8c-68d0-4aef-93a1-2f3e7c10d67e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.849942] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efab5dea-133d-493e-8db8-6c0b0cb2de56 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.880327] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5191baee-929f-4756-99f7-d29ca4371390 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.887840] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90fb8a40-6885-4d29-899b-3b3686294b3b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.902243] env[63293]: DEBUG nova.compute.provider_tree [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.406030] env[63293]: DEBUG nova.scheduler.client.report [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.910962] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.911641] env[63293]: ERROR nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 52e67338-7b66-47ff-8bea-0783bb5999a1, please check neutron logs for more information. [ 685.911641] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Traceback (most recent call last): [ 685.911641] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 685.911641] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] self.driver.spawn(context, instance, image_meta, [ 685.911641] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 685.911641] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.911641] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.911641] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] vm_ref = self.build_virtual_machine(instance, [ 685.911641] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.911641] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.911641] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] for vif in network_info: [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] return self._sync_wrapper(fn, *args, **kwargs) [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] self.wait() [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] self[:] = self._gt.wait() [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] return self._exit_event.wait() [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] current.throw(*self._exc) [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.911956] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] result = function(*args, **kwargs) [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] return func(*args, **kwargs) [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] raise e [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] nwinfo = self.network_api.allocate_for_instance( [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] created_port_ids = self._update_ports_for_instance( [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] with excutils.save_and_reraise_exception(): [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] self.force_reraise() [ 685.912283] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.912679] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] raise self.value [ 685.912679] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.912679] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] updated_port = self._update_port( [ 685.912679] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.912679] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] _ensure_no_port_binding_failure(port) [ 685.912679] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.912679] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] raise exception.PortBindingFailed(port_id=port['id']) [ 685.912679] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] nova.exception.PortBindingFailed: Binding failed for port 52e67338-7b66-47ff-8bea-0783bb5999a1, please check neutron logs for more information. [ 685.912679] env[63293]: ERROR nova.compute.manager [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] [ 685.912679] env[63293]: DEBUG nova.compute.utils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Binding failed for port 52e67338-7b66-47ff-8bea-0783bb5999a1, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 685.913573] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.591s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.913767] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.915682] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.952s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.920604] env[63293]: DEBUG nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Build of instance b2c1e806-d0ae-455f-95e9-405a9557697f was re-scheduled: Binding failed for port 52e67338-7b66-47ff-8bea-0783bb5999a1, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 685.922031] env[63293]: DEBUG nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 685.922031] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquiring lock "refresh_cache-b2c1e806-d0ae-455f-95e9-405a9557697f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.922031] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Acquired lock "refresh_cache-b2c1e806-d0ae-455f-95e9-405a9557697f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.922031] env[63293]: DEBUG nova.network.neutron [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 685.940964] env[63293]: INFO nova.scheduler.client.report [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Deleted allocations for instance b2055187-edd8-4a8c-9811-183cf7ba3b39 [ 686.443321] env[63293]: DEBUG nova.network.neutron [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.448320] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9392ce3a-9865-49c8-b6b7-d48eb0a36511 tempest-ServerDiagnosticsV248Test-511862286 tempest-ServerDiagnosticsV248Test-511862286-project-member] Lock "b2055187-edd8-4a8c-9811-183cf7ba3b39" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.838s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.537316] env[63293]: DEBUG nova.network.neutron [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.836435] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28bf38ea-da95-410c-a63f-e18539c0862b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.845091] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf85cd20-1996-40d1-988b-ae127c872359 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.874097] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87da2ac-6aa7-4a11-a6de-10afb19c3e7c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.881160] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2638b842-14b8-4b63-8fa3-337bb9957039 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.897719] env[63293]: DEBUG nova.compute.provider_tree [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.040610] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Releasing lock "refresh_cache-b2c1e806-d0ae-455f-95e9-405a9557697f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.041308] env[63293]: DEBUG nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 687.041683] env[63293]: DEBUG nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.042019] env[63293]: DEBUG nova.network.neutron [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.056956] env[63293]: DEBUG nova.network.neutron [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.403018] env[63293]: DEBUG nova.scheduler.client.report [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.561058] env[63293]: DEBUG nova.network.neutron [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.906588] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.991s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.907222] env[63293]: ERROR nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6683eea7-366f-4165-846e-f1103c0a0264, please check neutron logs for more information. [ 687.907222] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Traceback (most recent call last): [ 687.907222] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.907222] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] self.driver.spawn(context, instance, image_meta, [ 687.907222] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 687.907222] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.907222] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.907222] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] vm_ref = self.build_virtual_machine(instance, [ 687.907222] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.907222] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.907222] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] for vif in network_info: [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] return self._sync_wrapper(fn, *args, **kwargs) [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] self.wait() [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] self[:] = self._gt.wait() [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] return self._exit_event.wait() [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] result = hub.switch() [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 687.907583] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] return self.greenlet.switch() [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] result = function(*args, **kwargs) [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] return func(*args, **kwargs) [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] raise e [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] nwinfo = self.network_api.allocate_for_instance( [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] created_port_ids = self._update_ports_for_instance( [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] with excutils.save_and_reraise_exception(): [ 687.908030] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.908408] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] self.force_reraise() [ 687.908408] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.908408] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] raise self.value [ 687.908408] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.908408] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] updated_port = self._update_port( [ 687.908408] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.908408] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] _ensure_no_port_binding_failure(port) [ 687.908408] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.908408] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] raise exception.PortBindingFailed(port_id=port['id']) [ 687.908408] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] nova.exception.PortBindingFailed: Binding failed for port 6683eea7-366f-4165-846e-f1103c0a0264, please check neutron logs for more information. [ 687.908408] env[63293]: ERROR nova.compute.manager [instance: f5c47736-381f-4979-b804-494ca8cf16a1] [ 687.908735] env[63293]: DEBUG nova.compute.utils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Binding failed for port 6683eea7-366f-4165-846e-f1103c0a0264, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 687.909248] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.024s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.911084] env[63293]: INFO nova.compute.claims [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.913659] env[63293]: DEBUG nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Build of instance f5c47736-381f-4979-b804-494ca8cf16a1 was re-scheduled: Binding failed for port 6683eea7-366f-4165-846e-f1103c0a0264, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 687.914112] env[63293]: DEBUG nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 687.914338] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-f5c47736-381f-4979-b804-494ca8cf16a1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.914481] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-f5c47736-381f-4979-b804-494ca8cf16a1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.914637] env[63293]: DEBUG nova.network.neutron [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.063460] env[63293]: INFO nova.compute.manager [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] [instance: b2c1e806-d0ae-455f-95e9-405a9557697f] Took 1.02 seconds to deallocate network for instance. [ 688.466911] env[63293]: DEBUG nova.network.neutron [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.562054] env[63293]: DEBUG nova.network.neutron [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.062862] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-f5c47736-381f-4979-b804-494ca8cf16a1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.063124] env[63293]: DEBUG nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 689.063309] env[63293]: DEBUG nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.063477] env[63293]: DEBUG nova.network.neutron [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.083016] env[63293]: DEBUG nova.network.neutron [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.099392] env[63293]: INFO nova.scheduler.client.report [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Deleted allocations for instance b2c1e806-d0ae-455f-95e9-405a9557697f [ 689.332849] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12de74cb-da29-4a31-9186-bf5d2c8e65ab {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.340808] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d90454-8eab-433f-a92c-b6b1fa9202b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.371509] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d9f6e8-e983-4b19-9339-4b481c00fe3d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.379971] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3e11fa-0eba-4ae8-b74a-2083010e39f1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.395971] env[63293]: DEBUG nova.compute.provider_tree [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.584980] env[63293]: DEBUG nova.network.neutron [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.612723] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2e447c29-fa2c-4426-9cbf-dc0f5f1bd1a7 tempest-ListImageFiltersTestJSON-39817542 tempest-ListImageFiltersTestJSON-39817542-project-member] Lock "b2c1e806-d0ae-455f-95e9-405a9557697f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.605s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.897420] env[63293]: DEBUG nova.scheduler.client.report [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.090033] env[63293]: INFO nova.compute.manager [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f5c47736-381f-4979-b804-494ca8cf16a1] Took 1.03 seconds to deallocate network for instance. [ 690.118376] env[63293]: DEBUG nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 690.405129] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.405675] env[63293]: DEBUG nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 690.408209] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.402s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.643884] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.912719] env[63293]: DEBUG nova.compute.utils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.924923] env[63293]: DEBUG nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 690.924923] env[63293]: DEBUG nova.network.neutron [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 691.005240] env[63293]: DEBUG nova.policy [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '79b653c0cb0145d3812bc21722bf2d5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b50c7b307c340c1bc6643b94dc12587', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 691.125106] env[63293]: INFO nova.scheduler.client.report [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Deleted allocations for instance f5c47736-381f-4979-b804-494ca8cf16a1 [ 691.347284] env[63293]: DEBUG nova.network.neutron [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Successfully created port: ca939f80-43b1-4669-9070-f2300e069e4e {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.376356] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6aa109-54f8-4052-8d95-4eef32a123f9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.384322] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-706778ae-27db-4012-a099-6dccae8c241e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.418261] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb20f32-15eb-4bb0-8d58-736262094ff8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.425842] env[63293]: DEBUG nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 691.429261] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1febaa47-cc81-4a64-a8d0-cf498c8264cf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.443411] env[63293]: DEBUG nova.compute.provider_tree [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.635387] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8c36f9a-cb91-479c-beb7-57463ce7b6be tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "f5c47736-381f-4979-b804-494ca8cf16a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.041s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.949018] env[63293]: DEBUG nova.scheduler.client.report [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.139938] env[63293]: DEBUG nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.167534] env[63293]: DEBUG nova.compute.manager [req-d53bb3aa-6e78-4549-86af-595c87f7b47b req-620a291a-059b-4d75-87b7-531dcca4e8cc service nova] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Received event network-changed-ca939f80-43b1-4669-9070-f2300e069e4e {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 692.167711] env[63293]: DEBUG nova.compute.manager [req-d53bb3aa-6e78-4549-86af-595c87f7b47b req-620a291a-059b-4d75-87b7-531dcca4e8cc service nova] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Refreshing instance network info cache due to event network-changed-ca939f80-43b1-4669-9070-f2300e069e4e. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 692.167972] env[63293]: DEBUG oslo_concurrency.lockutils [req-d53bb3aa-6e78-4549-86af-595c87f7b47b req-620a291a-059b-4d75-87b7-531dcca4e8cc service nova] Acquiring lock "refresh_cache-92208bca-bbac-48fa-83db-3f721e6c80c2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.168524] env[63293]: DEBUG oslo_concurrency.lockutils [req-d53bb3aa-6e78-4549-86af-595c87f7b47b req-620a291a-059b-4d75-87b7-531dcca4e8cc service nova] Acquired lock "refresh_cache-92208bca-bbac-48fa-83db-3f721e6c80c2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.168728] env[63293]: DEBUG nova.network.neutron [req-d53bb3aa-6e78-4549-86af-595c87f7b47b req-620a291a-059b-4d75-87b7-531dcca4e8cc service nova] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Refreshing network info cache for port ca939f80-43b1-4669-9070-f2300e069e4e {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 692.379382] env[63293]: ERROR nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ca939f80-43b1-4669-9070-f2300e069e4e, please check neutron logs for more information. [ 692.379382] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 692.379382] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.379382] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 692.379382] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.379382] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 692.379382] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.379382] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 692.379382] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.379382] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 692.379382] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.379382] env[63293]: ERROR nova.compute.manager raise self.value [ 692.379382] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.379382] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 692.379382] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.379382] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 692.379820] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.379820] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 692.379820] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ca939f80-43b1-4669-9070-f2300e069e4e, please check neutron logs for more information. [ 692.379820] env[63293]: ERROR nova.compute.manager [ 692.379820] env[63293]: Traceback (most recent call last): [ 692.379820] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 692.379820] env[63293]: listener.cb(fileno) [ 692.379820] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.379820] env[63293]: result = function(*args, **kwargs) [ 692.379820] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.379820] env[63293]: return func(*args, **kwargs) [ 692.379820] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.379820] env[63293]: raise e [ 692.379820] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.379820] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 692.379820] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.379820] env[63293]: created_port_ids = self._update_ports_for_instance( [ 692.379820] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.379820] env[63293]: with excutils.save_and_reraise_exception(): [ 692.379820] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.379820] env[63293]: self.force_reraise() [ 692.379820] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.379820] env[63293]: raise self.value [ 692.379820] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.379820] env[63293]: updated_port = self._update_port( [ 692.379820] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.379820] env[63293]: _ensure_no_port_binding_failure(port) [ 692.379820] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.379820] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 692.380574] env[63293]: nova.exception.PortBindingFailed: Binding failed for port ca939f80-43b1-4669-9070-f2300e069e4e, please check neutron logs for more information. [ 692.380574] env[63293]: Removing descriptor: 15 [ 692.441459] env[63293]: DEBUG nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.455469] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.047s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.456128] env[63293]: ERROR nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 651b6bbf-b272-429d-8a2f-a0f7eae93e81, please check neutron logs for more information. [ 692.456128] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Traceback (most recent call last): [ 692.456128] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.456128] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] self.driver.spawn(context, instance, image_meta, [ 692.456128] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 692.456128] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.456128] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.456128] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] vm_ref = self.build_virtual_machine(instance, [ 692.456128] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.456128] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.456128] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] for vif in network_info: [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] return self._sync_wrapper(fn, *args, **kwargs) [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] self.wait() [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] self[:] = self._gt.wait() [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] return self._exit_event.wait() [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] result = hub.switch() [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 692.457486] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] return self.greenlet.switch() [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] result = function(*args, **kwargs) [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] return func(*args, **kwargs) [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] raise e [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] nwinfo = self.network_api.allocate_for_instance( [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] created_port_ids = self._update_ports_for_instance( [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] with excutils.save_and_reraise_exception(): [ 692.457972] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.458756] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] self.force_reraise() [ 692.458756] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.458756] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] raise self.value [ 692.458756] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.458756] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] updated_port = self._update_port( [ 692.458756] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.458756] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] _ensure_no_port_binding_failure(port) [ 692.458756] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.458756] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] raise exception.PortBindingFailed(port_id=port['id']) [ 692.458756] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] nova.exception.PortBindingFailed: Binding failed for port 651b6bbf-b272-429d-8a2f-a0f7eae93e81, please check neutron logs for more information. [ 692.458756] env[63293]: ERROR nova.compute.manager [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] [ 692.459297] env[63293]: DEBUG nova.compute.utils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Binding failed for port 651b6bbf-b272-429d-8a2f-a0f7eae93e81, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 692.459297] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.072s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.460502] env[63293]: INFO nova.compute.claims [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.468712] env[63293]: DEBUG nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Build of instance 947f69fa-77cf-4c7d-ba62-37567318b969 was re-scheduled: Binding failed for port 651b6bbf-b272-429d-8a2f-a0f7eae93e81, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 692.470585] env[63293]: DEBUG nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 692.471552] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-947f69fa-77cf-4c7d-ba62-37567318b969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.471732] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-947f69fa-77cf-4c7d-ba62-37567318b969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.471895] env[63293]: DEBUG nova.network.neutron [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 692.479662] env[63293]: DEBUG nova.virt.hardware [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.479876] env[63293]: DEBUG nova.virt.hardware [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.480164] env[63293]: DEBUG nova.virt.hardware [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.480241] env[63293]: DEBUG nova.virt.hardware [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.480404] env[63293]: DEBUG nova.virt.hardware [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.480546] env[63293]: DEBUG nova.virt.hardware [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.480747] env[63293]: DEBUG nova.virt.hardware [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.480904] env[63293]: DEBUG nova.virt.hardware [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.481105] env[63293]: DEBUG nova.virt.hardware [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.481269] env[63293]: DEBUG nova.virt.hardware [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.481439] env[63293]: DEBUG nova.virt.hardware [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.482372] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5b4b0f-b22f-4607-855e-3725e8258f61 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.495022] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed794af-a9c7-49be-a21d-8225d3a822f2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.511845] env[63293]: ERROR nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ca939f80-43b1-4669-9070-f2300e069e4e, please check neutron logs for more information. [ 692.511845] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Traceback (most recent call last): [ 692.511845] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 692.511845] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] yield resources [ 692.511845] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.511845] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] self.driver.spawn(context, instance, image_meta, [ 692.511845] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 692.511845] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.511845] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.511845] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] vm_ref = self.build_virtual_machine(instance, [ 692.511845] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] for vif in network_info: [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] return self._sync_wrapper(fn, *args, **kwargs) [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] self.wait() [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] self[:] = self._gt.wait() [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] return self._exit_event.wait() [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 692.512222] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] current.throw(*self._exc) [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] result = function(*args, **kwargs) [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] return func(*args, **kwargs) [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] raise e [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] nwinfo = self.network_api.allocate_for_instance( [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] created_port_ids = self._update_ports_for_instance( [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] with excutils.save_and_reraise_exception(): [ 692.512535] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.512894] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] self.force_reraise() [ 692.512894] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.512894] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] raise self.value [ 692.512894] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.512894] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] updated_port = self._update_port( [ 692.512894] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.512894] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] _ensure_no_port_binding_failure(port) [ 692.512894] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.512894] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] raise exception.PortBindingFailed(port_id=port['id']) [ 692.512894] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] nova.exception.PortBindingFailed: Binding failed for port ca939f80-43b1-4669-9070-f2300e069e4e, please check neutron logs for more information. [ 692.512894] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] [ 692.512894] env[63293]: INFO nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Terminating instance [ 692.514543] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquiring lock "refresh_cache-92208bca-bbac-48fa-83db-3f721e6c80c2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.665617] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.696525] env[63293]: DEBUG nova.network.neutron [req-d53bb3aa-6e78-4549-86af-595c87f7b47b req-620a291a-059b-4d75-87b7-531dcca4e8cc service nova] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.792065] env[63293]: DEBUG nova.network.neutron [req-d53bb3aa-6e78-4549-86af-595c87f7b47b req-620a291a-059b-4d75-87b7-531dcca4e8cc service nova] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.997753] env[63293]: DEBUG nova.network.neutron [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.081253] env[63293]: DEBUG nova.network.neutron [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.295074] env[63293]: DEBUG oslo_concurrency.lockutils [req-d53bb3aa-6e78-4549-86af-595c87f7b47b req-620a291a-059b-4d75-87b7-531dcca4e8cc service nova] Releasing lock "refresh_cache-92208bca-bbac-48fa-83db-3f721e6c80c2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.295504] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquired lock "refresh_cache-92208bca-bbac-48fa-83db-3f721e6c80c2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.295702] env[63293]: DEBUG nova.network.neutron [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.583502] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-947f69fa-77cf-4c7d-ba62-37567318b969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.583696] env[63293]: DEBUG nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 693.583886] env[63293]: DEBUG nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 693.584100] env[63293]: DEBUG nova.network.neutron [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 693.613800] env[63293]: DEBUG nova.network.neutron [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.824765] env[63293]: DEBUG nova.network.neutron [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.872850] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302358b9-46b9-4cb8-b46a-521645275a41 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.882727] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77f229b-3331-4cc6-be03-5578f24bdf97 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.922729] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e7d2f1-67df-4f69-a5d8-0df521b3fea0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.926017] env[63293]: DEBUG nova.network.neutron [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.931954] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8c549c-b5d1-42c7-ab57-94dd69c9d3e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.945480] env[63293]: DEBUG nova.compute.provider_tree [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.118879] env[63293]: DEBUG nova.network.neutron [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.194725] env[63293]: DEBUG nova.compute.manager [req-3fb500f8-3d30-41f1-8374-381eadf45802 req-e4e8aca3-5aa8-4497-aa33-f12f3535cedd service nova] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Received event network-vif-deleted-ca939f80-43b1-4669-9070-f2300e069e4e {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.350532] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.350532] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.428546] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Releasing lock "refresh_cache-92208bca-bbac-48fa-83db-3f721e6c80c2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.429308] env[63293]: DEBUG nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.429308] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.432021] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83141537-5967-4a4a-a51d-5a83c9d366ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.437788] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295cc61f-b6db-4eb6-87c4-4def30591adf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.449632] env[63293]: DEBUG nova.scheduler.client.report [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.463618] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 92208bca-bbac-48fa-83db-3f721e6c80c2 could not be found. [ 694.463862] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.464150] env[63293]: INFO nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 694.464426] env[63293]: DEBUG oslo.service.loopingcall [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.465216] env[63293]: DEBUG nova.compute.manager [-] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.465352] env[63293]: DEBUG nova.network.neutron [-] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.479521] env[63293]: DEBUG nova.network.neutron [-] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.621676] env[63293]: INFO nova.compute.manager [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 947f69fa-77cf-4c7d-ba62-37567318b969] Took 1.04 seconds to deallocate network for instance. [ 694.954592] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.955127] env[63293]: DEBUG nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 694.957681] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.622s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.960929] env[63293]: INFO nova.compute.claims [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.982443] env[63293]: DEBUG nova.network.neutron [-] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.465758] env[63293]: DEBUG nova.compute.utils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.469145] env[63293]: DEBUG nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 695.469145] env[63293]: DEBUG nova.network.neutron [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 695.487427] env[63293]: INFO nova.compute.manager [-] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Took 1.02 seconds to deallocate network for instance. [ 695.489198] env[63293]: DEBUG nova.compute.claims [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 695.489368] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.535301] env[63293]: DEBUG nova.policy [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03a004f9d0764e60b3b3b76ea99fb0b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6285f4d2ba024795bb144c8ff5d583e2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 695.652917] env[63293]: INFO nova.scheduler.client.report [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted allocations for instance 947f69fa-77cf-4c7d-ba62-37567318b969 [ 695.819886] env[63293]: DEBUG nova.network.neutron [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Successfully created port: 67f48fe2-558f-4c91-8a40-fc8b5a6e0e97 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 695.969141] env[63293]: DEBUG nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 696.163463] env[63293]: DEBUG oslo_concurrency.lockutils [None req-583c3fe4-c521-441d-93ef-4216cc255ee3 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "947f69fa-77cf-4c7d-ba62-37567318b969" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.487s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.392704] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e97e137e-3877-40af-b29e-be600ba88ae0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.401204] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9aacd2-0539-4f68-890a-c5016ab5a8a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.435109] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32cc70d-56be-4aea-be66-86ae392ce60c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.444169] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ea0a72-b630-4ef3-b059-dab78c601e15 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.459938] env[63293]: DEBUG nova.compute.provider_tree [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.665978] env[63293]: DEBUG nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 696.723290] env[63293]: DEBUG nova.compute.manager [req-26ecde15-c9e4-4962-b28c-1432dee4317c req-890aa2a8-d8f8-43f4-9462-99d99ec64c22 service nova] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Received event network-changed-67f48fe2-558f-4c91-8a40-fc8b5a6e0e97 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.723477] env[63293]: DEBUG nova.compute.manager [req-26ecde15-c9e4-4962-b28c-1432dee4317c req-890aa2a8-d8f8-43f4-9462-99d99ec64c22 service nova] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Refreshing instance network info cache due to event network-changed-67f48fe2-558f-4c91-8a40-fc8b5a6e0e97. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 696.723683] env[63293]: DEBUG oslo_concurrency.lockutils [req-26ecde15-c9e4-4962-b28c-1432dee4317c req-890aa2a8-d8f8-43f4-9462-99d99ec64c22 service nova] Acquiring lock "refresh_cache-305e3a9c-27cf-47b8-8825-09e5abe220ee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.723819] env[63293]: DEBUG oslo_concurrency.lockutils [req-26ecde15-c9e4-4962-b28c-1432dee4317c req-890aa2a8-d8f8-43f4-9462-99d99ec64c22 service nova] Acquired lock "refresh_cache-305e3a9c-27cf-47b8-8825-09e5abe220ee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.723973] env[63293]: DEBUG nova.network.neutron [req-26ecde15-c9e4-4962-b28c-1432dee4317c req-890aa2a8-d8f8-43f4-9462-99d99ec64c22 service nova] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Refreshing network info cache for port 67f48fe2-558f-4c91-8a40-fc8b5a6e0e97 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 696.966165] env[63293]: DEBUG nova.scheduler.client.report [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.980027] env[63293]: DEBUG nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 697.004097] env[63293]: DEBUG nova.virt.hardware [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:34:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='4b961f1e-9328-4802-9f64-442d29070870',id=26,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-658356586',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 697.004347] env[63293]: DEBUG nova.virt.hardware [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 697.004528] env[63293]: DEBUG nova.virt.hardware [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 697.004927] env[63293]: DEBUG nova.virt.hardware [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 697.004927] env[63293]: DEBUG nova.virt.hardware [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 697.005068] env[63293]: DEBUG nova.virt.hardware [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 697.005385] env[63293]: DEBUG nova.virt.hardware [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 697.005561] env[63293]: DEBUG nova.virt.hardware [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 697.005730] env[63293]: DEBUG nova.virt.hardware [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 697.005893] env[63293]: DEBUG nova.virt.hardware [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 697.006075] env[63293]: DEBUG nova.virt.hardware [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 697.007192] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3209d09-1c2e-4e57-bcea-66651dac8ea9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.015509] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e781a3b-a510-4527-9f7f-ee44394b361b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.059037] env[63293]: ERROR nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 67f48fe2-558f-4c91-8a40-fc8b5a6e0e97, please check neutron logs for more information. [ 697.059037] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 697.059037] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.059037] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 697.059037] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.059037] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 697.059037] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.059037] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 697.059037] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.059037] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 697.059037] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.059037] env[63293]: ERROR nova.compute.manager raise self.value [ 697.059037] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.059037] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 697.059037] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.059037] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 697.059562] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.059562] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 697.059562] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 67f48fe2-558f-4c91-8a40-fc8b5a6e0e97, please check neutron logs for more information. [ 697.059562] env[63293]: ERROR nova.compute.manager [ 697.059562] env[63293]: Traceback (most recent call last): [ 697.059562] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 697.059562] env[63293]: listener.cb(fileno) [ 697.059562] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.059562] env[63293]: result = function(*args, **kwargs) [ 697.059562] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 697.059562] env[63293]: return func(*args, **kwargs) [ 697.059562] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.059562] env[63293]: raise e [ 697.059562] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.059562] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 697.059562] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.059562] env[63293]: created_port_ids = self._update_ports_for_instance( [ 697.059562] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.059562] env[63293]: with excutils.save_and_reraise_exception(): [ 697.059562] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.059562] env[63293]: self.force_reraise() [ 697.059562] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.059562] env[63293]: raise self.value [ 697.059562] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.059562] env[63293]: updated_port = self._update_port( [ 697.059562] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.059562] env[63293]: _ensure_no_port_binding_failure(port) [ 697.059562] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.059562] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 697.060602] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 67f48fe2-558f-4c91-8a40-fc8b5a6e0e97, please check neutron logs for more information. [ 697.060602] env[63293]: Removing descriptor: 15 [ 697.060602] env[63293]: ERROR nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 67f48fe2-558f-4c91-8a40-fc8b5a6e0e97, please check neutron logs for more information. [ 697.060602] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Traceback (most recent call last): [ 697.060602] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 697.060602] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] yield resources [ 697.060602] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 697.060602] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] self.driver.spawn(context, instance, image_meta, [ 697.060602] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 697.060602] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.060602] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.060602] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] vm_ref = self.build_virtual_machine(instance, [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] for vif in network_info: [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] return self._sync_wrapper(fn, *args, **kwargs) [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] self.wait() [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] self[:] = self._gt.wait() [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] return self._exit_event.wait() [ 697.061070] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] result = hub.switch() [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] return self.greenlet.switch() [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] result = function(*args, **kwargs) [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] return func(*args, **kwargs) [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] raise e [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] nwinfo = self.network_api.allocate_for_instance( [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.061470] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] created_port_ids = self._update_ports_for_instance( [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] with excutils.save_and_reraise_exception(): [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] self.force_reraise() [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] raise self.value [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] updated_port = self._update_port( [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] _ensure_no_port_binding_failure(port) [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.061869] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] raise exception.PortBindingFailed(port_id=port['id']) [ 697.062239] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] nova.exception.PortBindingFailed: Binding failed for port 67f48fe2-558f-4c91-8a40-fc8b5a6e0e97, please check neutron logs for more information. [ 697.062239] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] [ 697.062239] env[63293]: INFO nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Terminating instance [ 697.063102] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquiring lock "refresh_cache-305e3a9c-27cf-47b8-8825-09e5abe220ee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.190752] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.241927] env[63293]: DEBUG nova.network.neutron [req-26ecde15-c9e4-4962-b28c-1432dee4317c req-890aa2a8-d8f8-43f4-9462-99d99ec64c22 service nova] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.299410] env[63293]: DEBUG nova.network.neutron [req-26ecde15-c9e4-4962-b28c-1432dee4317c req-890aa2a8-d8f8-43f4-9462-99d99ec64c22 service nova] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.472129] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.472669] env[63293]: DEBUG nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.475212] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.187s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.476690] env[63293]: INFO nova.compute.claims [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.714086] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "33d71260-26f1-482c-b93b-5f8e27c545f3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.714324] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "33d71260-26f1-482c-b93b-5f8e27c545f3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.802054] env[63293]: DEBUG oslo_concurrency.lockutils [req-26ecde15-c9e4-4962-b28c-1432dee4317c req-890aa2a8-d8f8-43f4-9462-99d99ec64c22 service nova] Releasing lock "refresh_cache-305e3a9c-27cf-47b8-8825-09e5abe220ee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.802460] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquired lock "refresh_cache-305e3a9c-27cf-47b8-8825-09e5abe220ee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.802636] env[63293]: DEBUG nova.network.neutron [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.980890] env[63293]: DEBUG nova.compute.utils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 697.984285] env[63293]: DEBUG nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 697.984455] env[63293]: DEBUG nova.network.neutron [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 698.036251] env[63293]: DEBUG nova.policy [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dcd81fddd8f047c1aee797d57d52805a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ac7ded391a34622aacc6435d795e6eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 698.469995] env[63293]: DEBUG nova.network.neutron [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.485884] env[63293]: DEBUG nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.525602] env[63293]: DEBUG nova.network.neutron [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Successfully created port: 95b75e3a-aced-4da6-8f6f-628e8a758fa7 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.615328] env[63293]: DEBUG nova.network.neutron [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.907151] env[63293]: DEBUG nova.compute.manager [req-45ee2cb7-0b40-47df-94dc-640a8efb309b req-ff9257c3-39d6-4caf-956b-2b3d90149832 service nova] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Received event network-vif-deleted-67f48fe2-558f-4c91-8a40-fc8b5a6e0e97 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 698.959689] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70735896-81e1-4bfa-8d15-368a3352757a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.967875] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483ed363-1e9d-41cd-bdb0-3b49e9a76e1e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.005086] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e3203c-895b-4c46-b570-7a2884f449c2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.011036] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a25261-eb84-4589-be2e-b173f977df8c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.026438] env[63293]: DEBUG nova.compute.provider_tree [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.124023] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Releasing lock "refresh_cache-305e3a9c-27cf-47b8-8825-09e5abe220ee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.124023] env[63293]: DEBUG nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 699.124023] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 699.124023] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-439cffd4-bf21-486c-8543-d975cf190749 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.132623] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3908476d-44d5-492f-b162-18e0fc0fe9b9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.160166] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 305e3a9c-27cf-47b8-8825-09e5abe220ee could not be found. [ 699.160166] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 699.160166] env[63293]: INFO nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Took 0.04 seconds to destroy the instance on the hypervisor. [ 699.160166] env[63293]: DEBUG oslo.service.loopingcall [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 699.160166] env[63293]: DEBUG nova.compute.manager [-] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 699.160166] env[63293]: DEBUG nova.network.neutron [-] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 699.191082] env[63293]: DEBUG nova.network.neutron [-] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.508294] env[63293]: DEBUG nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.531359] env[63293]: DEBUG nova.scheduler.client.report [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.543849] env[63293]: DEBUG nova.virt.hardware [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.544401] env[63293]: DEBUG nova.virt.hardware [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.544573] env[63293]: DEBUG nova.virt.hardware [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.544799] env[63293]: DEBUG nova.virt.hardware [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.545013] env[63293]: DEBUG nova.virt.hardware [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.545214] env[63293]: DEBUG nova.virt.hardware [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.545462] env[63293]: DEBUG nova.virt.hardware [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.545658] env[63293]: DEBUG nova.virt.hardware [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.545884] env[63293]: DEBUG nova.virt.hardware [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.546095] env[63293]: DEBUG nova.virt.hardware [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.546309] env[63293]: DEBUG nova.virt.hardware [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.547342] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f771f157-7eda-4618-92d1-54a3ce91f416 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.557493] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff782ef1-b6bd-4fed-980c-6e6a24dc1277 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.608369] env[63293]: ERROR nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 95b75e3a-aced-4da6-8f6f-628e8a758fa7, please check neutron logs for more information. [ 699.608369] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 699.608369] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.608369] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 699.608369] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.608369] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 699.608369] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.608369] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 699.608369] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.608369] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 699.608369] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.608369] env[63293]: ERROR nova.compute.manager raise self.value [ 699.608369] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.608369] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 699.608369] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.608369] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 699.608974] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.608974] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 699.608974] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 95b75e3a-aced-4da6-8f6f-628e8a758fa7, please check neutron logs for more information. [ 699.608974] env[63293]: ERROR nova.compute.manager [ 699.608974] env[63293]: Traceback (most recent call last): [ 699.608974] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 699.608974] env[63293]: listener.cb(fileno) [ 699.608974] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.608974] env[63293]: result = function(*args, **kwargs) [ 699.608974] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.608974] env[63293]: return func(*args, **kwargs) [ 699.608974] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.608974] env[63293]: raise e [ 699.608974] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.608974] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 699.608974] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.608974] env[63293]: created_port_ids = self._update_ports_for_instance( [ 699.608974] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.608974] env[63293]: with excutils.save_and_reraise_exception(): [ 699.608974] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.608974] env[63293]: self.force_reraise() [ 699.608974] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.608974] env[63293]: raise self.value [ 699.608974] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.608974] env[63293]: updated_port = self._update_port( [ 699.608974] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.608974] env[63293]: _ensure_no_port_binding_failure(port) [ 699.608974] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.608974] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 699.609960] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 95b75e3a-aced-4da6-8f6f-628e8a758fa7, please check neutron logs for more information. [ 699.609960] env[63293]: Removing descriptor: 15 [ 699.609960] env[63293]: ERROR nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 95b75e3a-aced-4da6-8f6f-628e8a758fa7, please check neutron logs for more information. [ 699.609960] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Traceback (most recent call last): [ 699.609960] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.609960] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] yield resources [ 699.609960] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.609960] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] self.driver.spawn(context, instance, image_meta, [ 699.609960] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 699.609960] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.609960] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.609960] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] vm_ref = self.build_virtual_machine(instance, [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] for vif in network_info: [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] return self._sync_wrapper(fn, *args, **kwargs) [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] self.wait() [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] self[:] = self._gt.wait() [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] return self._exit_event.wait() [ 699.610349] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] result = hub.switch() [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] return self.greenlet.switch() [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] result = function(*args, **kwargs) [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] return func(*args, **kwargs) [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] raise e [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] nwinfo = self.network_api.allocate_for_instance( [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.610746] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] created_port_ids = self._update_ports_for_instance( [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] with excutils.save_and_reraise_exception(): [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] self.force_reraise() [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] raise self.value [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] updated_port = self._update_port( [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] _ensure_no_port_binding_failure(port) [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.611254] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] raise exception.PortBindingFailed(port_id=port['id']) [ 699.611589] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] nova.exception.PortBindingFailed: Binding failed for port 95b75e3a-aced-4da6-8f6f-628e8a758fa7, please check neutron logs for more information. [ 699.611589] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] [ 699.611589] env[63293]: INFO nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Terminating instance [ 699.614678] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Acquiring lock "refresh_cache-a63744c8-b9bf-4a89-9319-061d53de6b5f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.614847] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Acquired lock "refresh_cache-a63744c8-b9bf-4a89-9319-061d53de6b5f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.615019] env[63293]: DEBUG nova.network.neutron [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.695466] env[63293]: DEBUG nova.network.neutron [-] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.039426] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.039974] env[63293]: DEBUG nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 700.042803] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.879s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.044559] env[63293]: INFO nova.compute.claims [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.142624] env[63293]: DEBUG nova.network.neutron [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.198581] env[63293]: INFO nova.compute.manager [-] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Took 1.04 seconds to deallocate network for instance. [ 700.201128] env[63293]: DEBUG nova.compute.claims [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 700.201340] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.234106] env[63293]: DEBUG nova.network.neutron [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.551189] env[63293]: DEBUG nova.compute.utils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 700.554212] env[63293]: DEBUG nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 700.554388] env[63293]: DEBUG nova.network.neutron [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 700.596688] env[63293]: DEBUG nova.policy [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '44c02cc48e8a4a4f986910d2f19e5098', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86fe0268e83342e9bc638af11f6e73c4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 700.736841] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Releasing lock "refresh_cache-a63744c8-b9bf-4a89-9319-061d53de6b5f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.737367] env[63293]: DEBUG nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 700.737572] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 700.737969] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7cb4777b-d7e3-4bc0-a3d6-079e7db37f86 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.746482] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405ae847-4166-4292-8a93-8de33ab2194e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.773131] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a63744c8-b9bf-4a89-9319-061d53de6b5f could not be found. [ 700.773131] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 700.773309] env[63293]: INFO nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 700.773521] env[63293]: DEBUG oslo.service.loopingcall [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.773751] env[63293]: DEBUG nova.compute.manager [-] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.773843] env[63293]: DEBUG nova.network.neutron [-] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.788527] env[63293]: DEBUG nova.network.neutron [-] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.923214] env[63293]: DEBUG nova.network.neutron [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Successfully created port: 2117f165-5629-4dbc-8031-aa8d4898becb {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.941450] env[63293]: DEBUG nova.compute.manager [req-6a013a6d-8ede-4a3c-a0f4-6656c6122c16 req-33b0f2e3-a5f6-406a-b1e6-fd702055eec7 service nova] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Received event network-changed-95b75e3a-aced-4da6-8f6f-628e8a758fa7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.941662] env[63293]: DEBUG nova.compute.manager [req-6a013a6d-8ede-4a3c-a0f4-6656c6122c16 req-33b0f2e3-a5f6-406a-b1e6-fd702055eec7 service nova] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Refreshing instance network info cache due to event network-changed-95b75e3a-aced-4da6-8f6f-628e8a758fa7. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 700.941841] env[63293]: DEBUG oslo_concurrency.lockutils [req-6a013a6d-8ede-4a3c-a0f4-6656c6122c16 req-33b0f2e3-a5f6-406a-b1e6-fd702055eec7 service nova] Acquiring lock "refresh_cache-a63744c8-b9bf-4a89-9319-061d53de6b5f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.941976] env[63293]: DEBUG oslo_concurrency.lockutils [req-6a013a6d-8ede-4a3c-a0f4-6656c6122c16 req-33b0f2e3-a5f6-406a-b1e6-fd702055eec7 service nova] Acquired lock "refresh_cache-a63744c8-b9bf-4a89-9319-061d53de6b5f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.942363] env[63293]: DEBUG nova.network.neutron [req-6a013a6d-8ede-4a3c-a0f4-6656c6122c16 req-33b0f2e3-a5f6-406a-b1e6-fd702055eec7 service nova] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Refreshing network info cache for port 95b75e3a-aced-4da6-8f6f-628e8a758fa7 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 701.064683] env[63293]: DEBUG nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 701.291343] env[63293]: DEBUG nova.network.neutron [-] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.406426] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2901b50-9f15-41b1-84e1-4bfbf50c7098 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.414279] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1da62d4-d728-4af1-bd20-1763eab92c5d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.447904] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa61a5ac-b1f5-46bd-9dd5-5c5086bd2340 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.455788] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd62913e-e22a-429d-a115-d9927c0b3e9f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.469750] env[63293]: DEBUG nova.compute.provider_tree [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.475020] env[63293]: DEBUG nova.network.neutron [req-6a013a6d-8ede-4a3c-a0f4-6656c6122c16 req-33b0f2e3-a5f6-406a-b1e6-fd702055eec7 service nova] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.549741] env[63293]: DEBUG nova.network.neutron [req-6a013a6d-8ede-4a3c-a0f4-6656c6122c16 req-33b0f2e3-a5f6-406a-b1e6-fd702055eec7 service nova] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.798868] env[63293]: INFO nova.compute.manager [-] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Took 1.02 seconds to deallocate network for instance. [ 701.800339] env[63293]: DEBUG nova.compute.claims [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 701.800339] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.921606] env[63293]: ERROR nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2117f165-5629-4dbc-8031-aa8d4898becb, please check neutron logs for more information. [ 701.921606] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.921606] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.921606] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.921606] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.921606] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.921606] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.921606] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.921606] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.921606] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 701.921606] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.921606] env[63293]: ERROR nova.compute.manager raise self.value [ 701.921606] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.921606] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.921606] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.921606] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.922114] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.922114] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.922114] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2117f165-5629-4dbc-8031-aa8d4898becb, please check neutron logs for more information. [ 701.922114] env[63293]: ERROR nova.compute.manager [ 701.922114] env[63293]: Traceback (most recent call last): [ 701.922114] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.922114] env[63293]: listener.cb(fileno) [ 701.922114] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.922114] env[63293]: result = function(*args, **kwargs) [ 701.922114] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 701.922114] env[63293]: return func(*args, **kwargs) [ 701.922114] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.922114] env[63293]: raise e [ 701.922114] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.922114] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 701.922114] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.922114] env[63293]: created_port_ids = self._update_ports_for_instance( [ 701.922114] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.922114] env[63293]: with excutils.save_and_reraise_exception(): [ 701.922114] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.922114] env[63293]: self.force_reraise() [ 701.922114] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.922114] env[63293]: raise self.value [ 701.922114] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.922114] env[63293]: updated_port = self._update_port( [ 701.922114] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.922114] env[63293]: _ensure_no_port_binding_failure(port) [ 701.922114] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.922114] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.922975] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 2117f165-5629-4dbc-8031-aa8d4898becb, please check neutron logs for more information. [ 701.922975] env[63293]: Removing descriptor: 15 [ 701.975035] env[63293]: DEBUG nova.scheduler.client.report [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.053135] env[63293]: DEBUG oslo_concurrency.lockutils [req-6a013a6d-8ede-4a3c-a0f4-6656c6122c16 req-33b0f2e3-a5f6-406a-b1e6-fd702055eec7 service nova] Releasing lock "refresh_cache-a63744c8-b9bf-4a89-9319-061d53de6b5f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.053372] env[63293]: DEBUG nova.compute.manager [req-6a013a6d-8ede-4a3c-a0f4-6656c6122c16 req-33b0f2e3-a5f6-406a-b1e6-fd702055eec7 service nova] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Received event network-vif-deleted-95b75e3a-aced-4da6-8f6f-628e8a758fa7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.077594] env[63293]: DEBUG nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 702.102594] env[63293]: DEBUG nova.virt.hardware [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 702.102829] env[63293]: DEBUG nova.virt.hardware [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 702.103029] env[63293]: DEBUG nova.virt.hardware [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 702.103248] env[63293]: DEBUG nova.virt.hardware [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 702.103395] env[63293]: DEBUG nova.virt.hardware [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 702.103539] env[63293]: DEBUG nova.virt.hardware [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 702.103742] env[63293]: DEBUG nova.virt.hardware [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 702.103898] env[63293]: DEBUG nova.virt.hardware [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 702.104072] env[63293]: DEBUG nova.virt.hardware [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 702.104236] env[63293]: DEBUG nova.virt.hardware [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 702.104406] env[63293]: DEBUG nova.virt.hardware [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 702.105421] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2692e1b-f440-47eb-8cba-42a9fc630515 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.113342] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eff3c62-e014-48b0-a1fa-2a11e2bea417 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.128502] env[63293]: ERROR nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2117f165-5629-4dbc-8031-aa8d4898becb, please check neutron logs for more information. [ 702.128502] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Traceback (most recent call last): [ 702.128502] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 702.128502] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] yield resources [ 702.128502] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.128502] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] self.driver.spawn(context, instance, image_meta, [ 702.128502] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 702.128502] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.128502] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.128502] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] vm_ref = self.build_virtual_machine(instance, [ 702.128502] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] for vif in network_info: [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] return self._sync_wrapper(fn, *args, **kwargs) [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] self.wait() [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] self[:] = self._gt.wait() [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] return self._exit_event.wait() [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 702.128819] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] current.throw(*self._exc) [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] result = function(*args, **kwargs) [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] return func(*args, **kwargs) [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] raise e [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] nwinfo = self.network_api.allocate_for_instance( [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] created_port_ids = self._update_ports_for_instance( [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] with excutils.save_and_reraise_exception(): [ 702.129307] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.129629] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] self.force_reraise() [ 702.129629] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.129629] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] raise self.value [ 702.129629] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.129629] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] updated_port = self._update_port( [ 702.129629] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.129629] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] _ensure_no_port_binding_failure(port) [ 702.129629] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.129629] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] raise exception.PortBindingFailed(port_id=port['id']) [ 702.129629] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] nova.exception.PortBindingFailed: Binding failed for port 2117f165-5629-4dbc-8031-aa8d4898becb, please check neutron logs for more information. [ 702.129629] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] [ 702.129629] env[63293]: INFO nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Terminating instance [ 702.130882] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Acquiring lock "refresh_cache-8b4b041b-5d3f-479a-8241-c75cce650d1e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.131054] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Acquired lock "refresh_cache-8b4b041b-5d3f-479a-8241-c75cce650d1e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.131224] env[63293]: DEBUG nova.network.neutron [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.480759] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.481362] env[63293]: DEBUG nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 702.483895] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.546s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.648734] env[63293]: DEBUG nova.network.neutron [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.724503] env[63293]: DEBUG nova.network.neutron [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.988514] env[63293]: DEBUG nova.compute.utils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 702.993208] env[63293]: DEBUG nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 702.993381] env[63293]: DEBUG nova.network.neutron [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 703.003825] env[63293]: DEBUG nova.compute.manager [req-ae36d9cf-f80e-4116-95ef-ede9845fd6d1 req-fa95f8a5-8177-42b5-b06b-bb6371bd130a service nova] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Received event network-changed-2117f165-5629-4dbc-8031-aa8d4898becb {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.004073] env[63293]: DEBUG nova.compute.manager [req-ae36d9cf-f80e-4116-95ef-ede9845fd6d1 req-fa95f8a5-8177-42b5-b06b-bb6371bd130a service nova] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Refreshing instance network info cache due to event network-changed-2117f165-5629-4dbc-8031-aa8d4898becb. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 703.004270] env[63293]: DEBUG oslo_concurrency.lockutils [req-ae36d9cf-f80e-4116-95ef-ede9845fd6d1 req-fa95f8a5-8177-42b5-b06b-bb6371bd130a service nova] Acquiring lock "refresh_cache-8b4b041b-5d3f-479a-8241-c75cce650d1e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.043970] env[63293]: DEBUG nova.policy [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '79b653c0cb0145d3812bc21722bf2d5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b50c7b307c340c1bc6643b94dc12587', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 703.226559] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Releasing lock "refresh_cache-8b4b041b-5d3f-479a-8241-c75cce650d1e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.226990] env[63293]: DEBUG nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.227195] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.227553] env[63293]: DEBUG oslo_concurrency.lockutils [req-ae36d9cf-f80e-4116-95ef-ede9845fd6d1 req-fa95f8a5-8177-42b5-b06b-bb6371bd130a service nova] Acquired lock "refresh_cache-8b4b041b-5d3f-479a-8241-c75cce650d1e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.227687] env[63293]: DEBUG nova.network.neutron [req-ae36d9cf-f80e-4116-95ef-ede9845fd6d1 req-fa95f8a5-8177-42b5-b06b-bb6371bd130a service nova] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Refreshing network info cache for port 2117f165-5629-4dbc-8031-aa8d4898becb {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 703.228856] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c101b69f-7b7c-483d-b53f-d84b79545195 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.248526] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16907be2-16cd-48e9-a351-09ade62becaa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.275442] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8b4b041b-5d3f-479a-8241-c75cce650d1e could not be found. [ 703.278093] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 703.278093] env[63293]: INFO nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 703.278093] env[63293]: DEBUG oslo.service.loopingcall [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.278731] env[63293]: DEBUG nova.compute.manager [-] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.278834] env[63293]: DEBUG nova.network.neutron [-] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.294715] env[63293]: DEBUG nova.network.neutron [-] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.361447] env[63293]: DEBUG nova.network.neutron [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Successfully created port: cd2f776b-4862-4992-964a-4c3ced3420eb {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 703.404630] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e43eeaf-d15d-4bdd-9f61-bc867fd8083b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.411845] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fcb51d0-e4b8-454a-bd04-666a964489c9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.440568] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365e92ac-429b-46a5-9315-dd03e278a16e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.447999] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72c2add-01ec-45b3-897c-08b8877cb0d9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.462363] env[63293]: DEBUG nova.compute.provider_tree [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.493360] env[63293]: DEBUG nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 703.756116] env[63293]: DEBUG nova.network.neutron [req-ae36d9cf-f80e-4116-95ef-ede9845fd6d1 req-fa95f8a5-8177-42b5-b06b-bb6371bd130a service nova] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.796752] env[63293]: DEBUG nova.network.neutron [-] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.845273] env[63293]: DEBUG nova.network.neutron [req-ae36d9cf-f80e-4116-95ef-ede9845fd6d1 req-fa95f8a5-8177-42b5-b06b-bb6371bd130a service nova] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.964930] env[63293]: DEBUG nova.scheduler.client.report [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.302987] env[63293]: INFO nova.compute.manager [-] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Took 1.02 seconds to deallocate network for instance. [ 704.305588] env[63293]: DEBUG nova.compute.claims [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 704.306048] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.331133] env[63293]: ERROR nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cd2f776b-4862-4992-964a-4c3ced3420eb, please check neutron logs for more information. [ 704.331133] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 704.331133] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.331133] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 704.331133] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.331133] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 704.331133] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.331133] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 704.331133] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.331133] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 704.331133] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.331133] env[63293]: ERROR nova.compute.manager raise self.value [ 704.331133] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.331133] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 704.331133] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.331133] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 704.331843] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.331843] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 704.331843] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cd2f776b-4862-4992-964a-4c3ced3420eb, please check neutron logs for more information. [ 704.331843] env[63293]: ERROR nova.compute.manager [ 704.331843] env[63293]: Traceback (most recent call last): [ 704.331843] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 704.331843] env[63293]: listener.cb(fileno) [ 704.331843] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.331843] env[63293]: result = function(*args, **kwargs) [ 704.331843] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.331843] env[63293]: return func(*args, **kwargs) [ 704.331843] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.331843] env[63293]: raise e [ 704.331843] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.331843] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 704.331843] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.331843] env[63293]: created_port_ids = self._update_ports_for_instance( [ 704.331843] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.331843] env[63293]: with excutils.save_and_reraise_exception(): [ 704.331843] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.331843] env[63293]: self.force_reraise() [ 704.331843] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.331843] env[63293]: raise self.value [ 704.331843] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.331843] env[63293]: updated_port = self._update_port( [ 704.331843] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.331843] env[63293]: _ensure_no_port_binding_failure(port) [ 704.331843] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.331843] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 704.333111] env[63293]: nova.exception.PortBindingFailed: Binding failed for port cd2f776b-4862-4992-964a-4c3ced3420eb, please check neutron logs for more information. [ 704.333111] env[63293]: Removing descriptor: 15 [ 704.349381] env[63293]: DEBUG oslo_concurrency.lockutils [req-ae36d9cf-f80e-4116-95ef-ede9845fd6d1 req-fa95f8a5-8177-42b5-b06b-bb6371bd130a service nova] Releasing lock "refresh_cache-8b4b041b-5d3f-479a-8241-c75cce650d1e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.349460] env[63293]: DEBUG nova.compute.manager [req-ae36d9cf-f80e-4116-95ef-ede9845fd6d1 req-fa95f8a5-8177-42b5-b06b-bb6371bd130a service nova] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Received event network-vif-deleted-2117f165-5629-4dbc-8031-aa8d4898becb {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.469918] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.986s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.470329] env[63293]: ERROR nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0, please check neutron logs for more information. [ 704.470329] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Traceback (most recent call last): [ 704.470329] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.470329] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] self.driver.spawn(context, instance, image_meta, [ 704.470329] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 704.470329] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.470329] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.470329] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] vm_ref = self.build_virtual_machine(instance, [ 704.470329] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.470329] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.470329] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] for vif in network_info: [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] return self._sync_wrapper(fn, *args, **kwargs) [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] self.wait() [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] self[:] = self._gt.wait() [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] return self._exit_event.wait() [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] result = hub.switch() [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 704.470817] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] return self.greenlet.switch() [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] result = function(*args, **kwargs) [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] return func(*args, **kwargs) [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] raise e [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] nwinfo = self.network_api.allocate_for_instance( [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] created_port_ids = self._update_ports_for_instance( [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] with excutils.save_and_reraise_exception(): [ 704.471242] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.471847] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] self.force_reraise() [ 704.471847] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.471847] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] raise self.value [ 704.471847] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.471847] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] updated_port = self._update_port( [ 704.471847] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.471847] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] _ensure_no_port_binding_failure(port) [ 704.471847] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.471847] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] raise exception.PortBindingFailed(port_id=port['id']) [ 704.471847] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] nova.exception.PortBindingFailed: Binding failed for port ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0, please check neutron logs for more information. [ 704.471847] env[63293]: ERROR nova.compute.manager [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] [ 704.472365] env[63293]: DEBUG nova.compute.utils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Binding failed for port ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 704.472365] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.781s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.473687] env[63293]: INFO nova.compute.claims [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.476214] env[63293]: DEBUG nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Build of instance dc9ae29a-a61b-4402-a16f-460cf74bc10c was re-scheduled: Binding failed for port ecc7b5a6-2dd0-4a51-8864-e76c0555c5b0, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 704.477028] env[63293]: DEBUG nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 704.477028] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Acquiring lock "refresh_cache-dc9ae29a-a61b-4402-a16f-460cf74bc10c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.477028] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Acquired lock "refresh_cache-dc9ae29a-a61b-4402-a16f-460cf74bc10c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.477200] env[63293]: DEBUG nova.network.neutron [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 704.502786] env[63293]: DEBUG nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 704.530414] env[63293]: DEBUG nova.virt.hardware [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 704.530739] env[63293]: DEBUG nova.virt.hardware [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 704.531019] env[63293]: DEBUG nova.virt.hardware [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 704.531297] env[63293]: DEBUG nova.virt.hardware [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 704.531684] env[63293]: DEBUG nova.virt.hardware [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 704.531812] env[63293]: DEBUG nova.virt.hardware [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 704.532153] env[63293]: DEBUG nova.virt.hardware [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 704.532394] env[63293]: DEBUG nova.virt.hardware [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 704.532643] env[63293]: DEBUG nova.virt.hardware [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 704.532893] env[63293]: DEBUG nova.virt.hardware [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 704.533168] env[63293]: DEBUG nova.virt.hardware [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 704.534537] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338713dc-807e-43c3-ae89-b7b954ecadaf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.544242] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cbaf514-9d80-42f7-8755-b80ed708b010 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.558887] env[63293]: ERROR nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cd2f776b-4862-4992-964a-4c3ced3420eb, please check neutron logs for more information. [ 704.558887] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Traceback (most recent call last): [ 704.558887] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 704.558887] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] yield resources [ 704.558887] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.558887] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] self.driver.spawn(context, instance, image_meta, [ 704.558887] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 704.558887] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.558887] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.558887] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] vm_ref = self.build_virtual_machine(instance, [ 704.558887] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] for vif in network_info: [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] return self._sync_wrapper(fn, *args, **kwargs) [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] self.wait() [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] self[:] = self._gt.wait() [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] return self._exit_event.wait() [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 704.559291] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] current.throw(*self._exc) [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] result = function(*args, **kwargs) [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] return func(*args, **kwargs) [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] raise e [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] nwinfo = self.network_api.allocate_for_instance( [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] created_port_ids = self._update_ports_for_instance( [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] with excutils.save_and_reraise_exception(): [ 704.559685] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.560088] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] self.force_reraise() [ 704.560088] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.560088] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] raise self.value [ 704.560088] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.560088] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] updated_port = self._update_port( [ 704.560088] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.560088] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] _ensure_no_port_binding_failure(port) [ 704.560088] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.560088] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] raise exception.PortBindingFailed(port_id=port['id']) [ 704.560088] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] nova.exception.PortBindingFailed: Binding failed for port cd2f776b-4862-4992-964a-4c3ced3420eb, please check neutron logs for more information. [ 704.560088] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] [ 704.560088] env[63293]: INFO nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Terminating instance [ 704.561134] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquiring lock "refresh_cache-8c2439e1-26ce-45e5-9179-164c0dbbbf76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.561293] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquired lock "refresh_cache-8c2439e1-26ce-45e5-9179-164c0dbbbf76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.561455] env[63293]: DEBUG nova.network.neutron [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.003988] env[63293]: DEBUG nova.network.neutron [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.037289] env[63293]: DEBUG nova.compute.manager [req-d1288c40-93e0-4e6e-a7dd-28b49d5cd5e1 req-a440f4ea-1725-4787-a4ee-3a540eaa4662 service nova] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Received event network-changed-cd2f776b-4862-4992-964a-4c3ced3420eb {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.037476] env[63293]: DEBUG nova.compute.manager [req-d1288c40-93e0-4e6e-a7dd-28b49d5cd5e1 req-a440f4ea-1725-4787-a4ee-3a540eaa4662 service nova] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Refreshing instance network info cache due to event network-changed-cd2f776b-4862-4992-964a-4c3ced3420eb. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 705.037700] env[63293]: DEBUG oslo_concurrency.lockutils [req-d1288c40-93e0-4e6e-a7dd-28b49d5cd5e1 req-a440f4ea-1725-4787-a4ee-3a540eaa4662 service nova] Acquiring lock "refresh_cache-8c2439e1-26ce-45e5-9179-164c0dbbbf76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.081801] env[63293]: DEBUG nova.network.neutron [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.099642] env[63293]: DEBUG nova.network.neutron [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.174989] env[63293]: DEBUG nova.network.neutron [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.603029] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Releasing lock "refresh_cache-dc9ae29a-a61b-4402-a16f-460cf74bc10c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.603029] env[63293]: DEBUG nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 705.603029] env[63293]: DEBUG nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.603029] env[63293]: DEBUG nova.network.neutron [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.617306] env[63293]: DEBUG nova.network.neutron [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.678077] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Releasing lock "refresh_cache-8c2439e1-26ce-45e5-9179-164c0dbbbf76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.678507] env[63293]: DEBUG nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 705.678695] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.679221] env[63293]: DEBUG oslo_concurrency.lockutils [req-d1288c40-93e0-4e6e-a7dd-28b49d5cd5e1 req-a440f4ea-1725-4787-a4ee-3a540eaa4662 service nova] Acquired lock "refresh_cache-8c2439e1-26ce-45e5-9179-164c0dbbbf76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.679406] env[63293]: DEBUG nova.network.neutron [req-d1288c40-93e0-4e6e-a7dd-28b49d5cd5e1 req-a440f4ea-1725-4787-a4ee-3a540eaa4662 service nova] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Refreshing network info cache for port cd2f776b-4862-4992-964a-4c3ced3420eb {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 705.680408] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0659821-764a-4d09-9ce5-cdb2ce2e4e7a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.690226] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7edf9886-6141-469b-a7b5-56beb77187e3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.714839] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8c2439e1-26ce-45e5-9179-164c0dbbbf76 could not be found. [ 705.715613] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 705.715613] env[63293]: INFO nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Took 0.04 seconds to destroy the instance on the hypervisor. [ 705.715613] env[63293]: DEBUG oslo.service.loopingcall [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.718039] env[63293]: DEBUG nova.compute.manager [-] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.718120] env[63293]: DEBUG nova.network.neutron [-] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.737661] env[63293]: DEBUG nova.network.neutron [-] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.846370] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3c8843-ada7-4a31-bbce-c97fe71a2ce4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.853777] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0f215f-54aa-4580-8f73-9a738654c4ad {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.883808] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8209964b-6d3d-4e7e-b44d-7c317c50b9ef {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.890883] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ed20ef-67fe-44dc-9362-229fd4d65b27 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.903551] env[63293]: DEBUG nova.compute.provider_tree [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.119952] env[63293]: DEBUG nova.network.neutron [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.200042] env[63293]: DEBUG nova.network.neutron [req-d1288c40-93e0-4e6e-a7dd-28b49d5cd5e1 req-a440f4ea-1725-4787-a4ee-3a540eaa4662 service nova] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.240535] env[63293]: DEBUG nova.network.neutron [-] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.289489] env[63293]: DEBUG nova.network.neutron [req-d1288c40-93e0-4e6e-a7dd-28b49d5cd5e1 req-a440f4ea-1725-4787-a4ee-3a540eaa4662 service nova] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.407045] env[63293]: DEBUG nova.scheduler.client.report [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.625024] env[63293]: INFO nova.compute.manager [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] [instance: dc9ae29a-a61b-4402-a16f-460cf74bc10c] Took 1.02 seconds to deallocate network for instance. [ 706.744894] env[63293]: INFO nova.compute.manager [-] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Took 1.03 seconds to deallocate network for instance. [ 706.747321] env[63293]: DEBUG nova.compute.claims [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 706.747500] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.791606] env[63293]: DEBUG oslo_concurrency.lockutils [req-d1288c40-93e0-4e6e-a7dd-28b49d5cd5e1 req-a440f4ea-1725-4787-a4ee-3a540eaa4662 service nova] Releasing lock "refresh_cache-8c2439e1-26ce-45e5-9179-164c0dbbbf76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.791801] env[63293]: DEBUG nova.compute.manager [req-d1288c40-93e0-4e6e-a7dd-28b49d5cd5e1 req-a440f4ea-1725-4787-a4ee-3a540eaa4662 service nova] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Received event network-vif-deleted-cd2f776b-4862-4992-964a-4c3ced3420eb {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 706.912110] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.912688] env[63293]: DEBUG nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.916043] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.272s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.917232] env[63293]: INFO nova.compute.claims [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.424760] env[63293]: DEBUG nova.compute.utils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 707.426993] env[63293]: DEBUG nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 707.427177] env[63293]: DEBUG nova.network.neutron [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 707.480758] env[63293]: DEBUG nova.policy [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9737df6638714a61978e8f81e306517c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2879022aafad4fd7968f22aa3eb0a85e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 707.661932] env[63293]: INFO nova.scheduler.client.report [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Deleted allocations for instance dc9ae29a-a61b-4402-a16f-460cf74bc10c [ 707.869104] env[63293]: DEBUG nova.network.neutron [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Successfully created port: 73e90e52-1f97-4b51-8dfa-3588700c63f3 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.929960] env[63293]: DEBUG nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 708.177629] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b09ecf71-b9dd-4a9f-8925-bdf379eee63d tempest-ServerGroupTestJSON-657322947 tempest-ServerGroupTestJSON-657322947-project-member] Lock "dc9ae29a-a61b-4402-a16f-460cf74bc10c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.206s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.302075] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2dd7bc-fe29-4db6-bec8-64289a3b6ed0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.311105] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f62f331-2dc6-4959-a850-94afbc114d57 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.346242] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1994c8-dc88-4cf8-bb0c-e4c45cc141c3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.354668] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0dbd2aa-6389-4a45-a560-6f7e6dd78277 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.373029] env[63293]: DEBUG nova.compute.provider_tree [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.680977] env[63293]: DEBUG nova.compute.manager [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 708.878030] env[63293]: DEBUG nova.scheduler.client.report [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.939917] env[63293]: DEBUG nova.compute.manager [req-1926457a-e219-4bb5-9ca2-581c39b146f8 req-3dbbf07b-278e-4a16-ab0b-46a444afbbca service nova] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Received event network-changed-73e90e52-1f97-4b51-8dfa-3588700c63f3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.940986] env[63293]: DEBUG nova.compute.manager [req-1926457a-e219-4bb5-9ca2-581c39b146f8 req-3dbbf07b-278e-4a16-ab0b-46a444afbbca service nova] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Refreshing instance network info cache due to event network-changed-73e90e52-1f97-4b51-8dfa-3588700c63f3. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 708.941364] env[63293]: DEBUG oslo_concurrency.lockutils [req-1926457a-e219-4bb5-9ca2-581c39b146f8 req-3dbbf07b-278e-4a16-ab0b-46a444afbbca service nova] Acquiring lock "refresh_cache-e0e08728-b6c3-4610-bc27-348c5594fefc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.942628] env[63293]: DEBUG oslo_concurrency.lockutils [req-1926457a-e219-4bb5-9ca2-581c39b146f8 req-3dbbf07b-278e-4a16-ab0b-46a444afbbca service nova] Acquired lock "refresh_cache-e0e08728-b6c3-4610-bc27-348c5594fefc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.942988] env[63293]: DEBUG nova.network.neutron [req-1926457a-e219-4bb5-9ca2-581c39b146f8 req-3dbbf07b-278e-4a16-ab0b-46a444afbbca service nova] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Refreshing network info cache for port 73e90e52-1f97-4b51-8dfa-3588700c63f3 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 708.945626] env[63293]: DEBUG nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 708.975542] env[63293]: ERROR nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 73e90e52-1f97-4b51-8dfa-3588700c63f3, please check neutron logs for more information. [ 708.975542] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 708.975542] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.975542] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 708.975542] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.975542] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 708.975542] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.975542] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 708.975542] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.975542] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 708.975542] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.975542] env[63293]: ERROR nova.compute.manager raise self.value [ 708.975542] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.975542] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 708.975542] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.975542] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 708.976129] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.976129] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 708.976129] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 73e90e52-1f97-4b51-8dfa-3588700c63f3, please check neutron logs for more information. [ 708.976129] env[63293]: ERROR nova.compute.manager [ 708.976129] env[63293]: Traceback (most recent call last): [ 708.976129] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 708.976129] env[63293]: listener.cb(fileno) [ 708.976129] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.976129] env[63293]: result = function(*args, **kwargs) [ 708.976129] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 708.976129] env[63293]: return func(*args, **kwargs) [ 708.976129] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.976129] env[63293]: raise e [ 708.976129] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.976129] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 708.976129] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.976129] env[63293]: created_port_ids = self._update_ports_for_instance( [ 708.976129] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.976129] env[63293]: with excutils.save_and_reraise_exception(): [ 708.976129] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.976129] env[63293]: self.force_reraise() [ 708.976129] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.976129] env[63293]: raise self.value [ 708.976129] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.976129] env[63293]: updated_port = self._update_port( [ 708.976129] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.976129] env[63293]: _ensure_no_port_binding_failure(port) [ 708.976129] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.976129] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 708.976904] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 73e90e52-1f97-4b51-8dfa-3588700c63f3, please check neutron logs for more information. [ 708.976904] env[63293]: Removing descriptor: 17 [ 708.978942] env[63293]: DEBUG nova.virt.hardware [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.979244] env[63293]: DEBUG nova.virt.hardware [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.979406] env[63293]: DEBUG nova.virt.hardware [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.979589] env[63293]: DEBUG nova.virt.hardware [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.979734] env[63293]: DEBUG nova.virt.hardware [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.979878] env[63293]: DEBUG nova.virt.hardware [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.980096] env[63293]: DEBUG nova.virt.hardware [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.980259] env[63293]: DEBUG nova.virt.hardware [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.980424] env[63293]: DEBUG nova.virt.hardware [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.980584] env[63293]: DEBUG nova.virt.hardware [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.980753] env[63293]: DEBUG nova.virt.hardware [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.981883] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3721acc4-d150-49f9-9549-a45f571d6c7a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.990395] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78790cde-4722-4f6c-9f1c-851ce8c39c79 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.005014] env[63293]: ERROR nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 73e90e52-1f97-4b51-8dfa-3588700c63f3, please check neutron logs for more information. [ 709.005014] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Traceback (most recent call last): [ 709.005014] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 709.005014] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] yield resources [ 709.005014] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 709.005014] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] self.driver.spawn(context, instance, image_meta, [ 709.005014] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 709.005014] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.005014] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.005014] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] vm_ref = self.build_virtual_machine(instance, [ 709.005014] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] for vif in network_info: [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] return self._sync_wrapper(fn, *args, **kwargs) [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] self.wait() [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] self[:] = self._gt.wait() [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] return self._exit_event.wait() [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 709.005464] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] current.throw(*self._exc) [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] result = function(*args, **kwargs) [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] return func(*args, **kwargs) [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] raise e [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] nwinfo = self.network_api.allocate_for_instance( [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] created_port_ids = self._update_ports_for_instance( [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] with excutils.save_and_reraise_exception(): [ 709.005835] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.006265] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] self.force_reraise() [ 709.006265] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.006265] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] raise self.value [ 709.006265] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.006265] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] updated_port = self._update_port( [ 709.006265] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.006265] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] _ensure_no_port_binding_failure(port) [ 709.006265] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.006265] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] raise exception.PortBindingFailed(port_id=port['id']) [ 709.006265] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] nova.exception.PortBindingFailed: Binding failed for port 73e90e52-1f97-4b51-8dfa-3588700c63f3, please check neutron logs for more information. [ 709.006265] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] [ 709.006265] env[63293]: INFO nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Terminating instance [ 709.007365] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Acquiring lock "refresh_cache-e0e08728-b6c3-4610-bc27-348c5594fefc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.203527] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.386872] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.387406] env[63293]: DEBUG nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 709.390503] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.725s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.392454] env[63293]: INFO nova.compute.claims [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.464800] env[63293]: DEBUG nova.network.neutron [req-1926457a-e219-4bb5-9ca2-581c39b146f8 req-3dbbf07b-278e-4a16-ab0b-46a444afbbca service nova] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.498705] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Acquiring lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.498705] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.557551] env[63293]: DEBUG nova.network.neutron [req-1926457a-e219-4bb5-9ca2-581c39b146f8 req-3dbbf07b-278e-4a16-ab0b-46a444afbbca service nova] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.899213] env[63293]: DEBUG nova.compute.utils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 709.909027] env[63293]: DEBUG nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 709.909027] env[63293]: DEBUG nova.network.neutron [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 709.951161] env[63293]: DEBUG nova.policy [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32e3c796e415413d8aaed9e0bc6f08c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9ce2a8ce367438b83ba557eb12f2a44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.063945] env[63293]: DEBUG oslo_concurrency.lockutils [req-1926457a-e219-4bb5-9ca2-581c39b146f8 req-3dbbf07b-278e-4a16-ab0b-46a444afbbca service nova] Releasing lock "refresh_cache-e0e08728-b6c3-4610-bc27-348c5594fefc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.064428] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Acquired lock "refresh_cache-e0e08728-b6c3-4610-bc27-348c5594fefc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.064614] env[63293]: DEBUG nova.network.neutron [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.284763] env[63293]: DEBUG nova.network.neutron [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Successfully created port: 11eb4a89-6521-4f4d-867e-977a50fd05ab {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.411678] env[63293]: DEBUG nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 710.597960] env[63293]: DEBUG nova.network.neutron [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.699391] env[63293]: DEBUG nova.network.neutron [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.837657] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2085e89a-3744-4e36-a9a0-13e787d5d8ab {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.848396] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742f226f-cb28-4afb-80a7-d0a59c14dce6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.882402] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271cc492-0cc5-4c26-8b8a-fa8b3c431ba6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.891554] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d2e4cd-c2aa-4773-a6f4-35324d782db0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.906342] env[63293]: DEBUG nova.compute.provider_tree [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.032622] env[63293]: DEBUG nova.compute.manager [req-dbdb46a0-82f3-4bb3-9aa2-93d029b2b1d2 req-72715003-6199-4452-baf9-373b1b9e573d service nova] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Received event network-vif-deleted-73e90e52-1f97-4b51-8dfa-3588700c63f3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 711.205212] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Releasing lock "refresh_cache-e0e08728-b6c3-4610-bc27-348c5594fefc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.205212] env[63293]: DEBUG nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 711.205212] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 711.205212] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c63723c-805f-4cfe-9b2a-168242ad8996 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.213776] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f9ccc1-ce04-423e-9a58-21010d4cfda4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.239098] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e0e08728-b6c3-4610-bc27-348c5594fefc could not be found. [ 711.239460] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 711.239697] env[63293]: INFO nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 711.239975] env[63293]: DEBUG oslo.service.loopingcall [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.240237] env[63293]: DEBUG nova.compute.manager [-] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.240440] env[63293]: DEBUG nova.network.neutron [-] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.267558] env[63293]: DEBUG nova.network.neutron [-] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.411451] env[63293]: DEBUG nova.scheduler.client.report [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.423115] env[63293]: DEBUG nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 711.453997] env[63293]: DEBUG nova.virt.hardware [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.454266] env[63293]: DEBUG nova.virt.hardware [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.454423] env[63293]: DEBUG nova.virt.hardware [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.454603] env[63293]: DEBUG nova.virt.hardware [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.454748] env[63293]: DEBUG nova.virt.hardware [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.454893] env[63293]: DEBUG nova.virt.hardware [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.455460] env[63293]: DEBUG nova.virt.hardware [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.455674] env[63293]: DEBUG nova.virt.hardware [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.455851] env[63293]: DEBUG nova.virt.hardware [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.456034] env[63293]: DEBUG nova.virt.hardware [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.456210] env[63293]: DEBUG nova.virt.hardware [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.457688] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddba16de-22ab-4618-a6e4-1c2ed0e3b71e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.469443] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235b0b87-5ef7-4b14-a774-8db8dd13536b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.713668] env[63293]: ERROR nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 11eb4a89-6521-4f4d-867e-977a50fd05ab, please check neutron logs for more information. [ 711.713668] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 711.713668] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.713668] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 711.713668] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.713668] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 711.713668] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.713668] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 711.713668] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.713668] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 711.713668] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.713668] env[63293]: ERROR nova.compute.manager raise self.value [ 711.713668] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.713668] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 711.713668] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.713668] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 711.714222] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.714222] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 711.714222] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 11eb4a89-6521-4f4d-867e-977a50fd05ab, please check neutron logs for more information. [ 711.714222] env[63293]: ERROR nova.compute.manager [ 711.714222] env[63293]: Traceback (most recent call last): [ 711.714222] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 711.714222] env[63293]: listener.cb(fileno) [ 711.714222] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.714222] env[63293]: result = function(*args, **kwargs) [ 711.714222] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 711.714222] env[63293]: return func(*args, **kwargs) [ 711.714222] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.714222] env[63293]: raise e [ 711.714222] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.714222] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 711.714222] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.714222] env[63293]: created_port_ids = self._update_ports_for_instance( [ 711.714222] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.714222] env[63293]: with excutils.save_and_reraise_exception(): [ 711.714222] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.714222] env[63293]: self.force_reraise() [ 711.714222] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.714222] env[63293]: raise self.value [ 711.714222] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.714222] env[63293]: updated_port = self._update_port( [ 711.714222] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.714222] env[63293]: _ensure_no_port_binding_failure(port) [ 711.714222] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.714222] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 711.715092] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 11eb4a89-6521-4f4d-867e-977a50fd05ab, please check neutron logs for more information. [ 711.715092] env[63293]: Removing descriptor: 17 [ 711.715092] env[63293]: ERROR nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 11eb4a89-6521-4f4d-867e-977a50fd05ab, please check neutron logs for more information. [ 711.715092] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Traceback (most recent call last): [ 711.715092] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 711.715092] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] yield resources [ 711.715092] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 711.715092] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] self.driver.spawn(context, instance, image_meta, [ 711.715092] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 711.715092] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.715092] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.715092] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] vm_ref = self.build_virtual_machine(instance, [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] for vif in network_info: [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] return self._sync_wrapper(fn, *args, **kwargs) [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] self.wait() [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] self[:] = self._gt.wait() [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] return self._exit_event.wait() [ 711.715487] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] result = hub.switch() [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] return self.greenlet.switch() [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] result = function(*args, **kwargs) [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] return func(*args, **kwargs) [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] raise e [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] nwinfo = self.network_api.allocate_for_instance( [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.715912] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] created_port_ids = self._update_ports_for_instance( [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] with excutils.save_and_reraise_exception(): [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] self.force_reraise() [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] raise self.value [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] updated_port = self._update_port( [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] _ensure_no_port_binding_failure(port) [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.716364] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] raise exception.PortBindingFailed(port_id=port['id']) [ 711.716765] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] nova.exception.PortBindingFailed: Binding failed for port 11eb4a89-6521-4f4d-867e-977a50fd05ab, please check neutron logs for more information. [ 711.716765] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] [ 711.716765] env[63293]: INFO nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Terminating instance [ 711.717708] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Acquiring lock "refresh_cache-2afc50fb-138c-4523-b086-db11026ae52f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.717906] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Acquired lock "refresh_cache-2afc50fb-138c-4523-b086-db11026ae52f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.718132] env[63293]: DEBUG nova.network.neutron [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.773359] env[63293]: DEBUG nova.network.neutron [-] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.923226] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.923766] env[63293]: DEBUG nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 711.926293] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.437s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.240233] env[63293]: DEBUG nova.network.neutron [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.278024] env[63293]: INFO nova.compute.manager [-] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Took 1.04 seconds to deallocate network for instance. [ 712.280245] env[63293]: DEBUG nova.compute.claims [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 712.280420] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.321205] env[63293]: DEBUG nova.network.neutron [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.431343] env[63293]: DEBUG nova.compute.utils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.435826] env[63293]: DEBUG nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 712.436045] env[63293]: DEBUG nova.network.neutron [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 712.491288] env[63293]: DEBUG nova.policy [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee523b37a5b141ecbf9c46ff255221d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25464bf6c45c4717854082b4218854d3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 712.824475] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Releasing lock "refresh_cache-2afc50fb-138c-4523-b086-db11026ae52f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.824992] env[63293]: DEBUG nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 712.825306] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.826545] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3f2b5a6-f464-4490-8ea5-a44b3b8a7f1e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.829519] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb93fd0-1e57-4769-b8fc-a959d388a7f8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.834437] env[63293]: DEBUG nova.network.neutron [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Successfully created port: 6c767097-c348-4bbf-b8c9-a1d2ab41926d {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.841851] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fcac80-a82e-42d0-b7a2-31636321d045 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.854124] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be30cf90-e8c6-4f17-8109-28015b19b8af {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.888393] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56afc54-1a1a-4476-97ae-02500cf944d9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.893613] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2afc50fb-138c-4523-b086-db11026ae52f could not be found. [ 712.893613] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 712.893613] env[63293]: INFO nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Took 0.07 seconds to destroy the instance on the hypervisor. [ 712.893613] env[63293]: DEBUG oslo.service.loopingcall [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 712.893613] env[63293]: DEBUG nova.compute.manager [-] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.893613] env[63293]: DEBUG nova.network.neutron [-] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.899529] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e4cac1-4658-4230-b386-d604046d290a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.914186] env[63293]: DEBUG nova.compute.provider_tree [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.931560] env[63293]: DEBUG nova.network.neutron [-] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.936560] env[63293]: DEBUG nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 713.059713] env[63293]: DEBUG nova.compute.manager [req-1a1f5748-e062-4289-a9c5-5e16f5fb6ada req-490ffeb9-196c-4cff-b5fc-131425e0b891 service nova] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Received event network-changed-11eb4a89-6521-4f4d-867e-977a50fd05ab {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 713.059895] env[63293]: DEBUG nova.compute.manager [req-1a1f5748-e062-4289-a9c5-5e16f5fb6ada req-490ffeb9-196c-4cff-b5fc-131425e0b891 service nova] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Refreshing instance network info cache due to event network-changed-11eb4a89-6521-4f4d-867e-977a50fd05ab. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 713.060810] env[63293]: DEBUG oslo_concurrency.lockutils [req-1a1f5748-e062-4289-a9c5-5e16f5fb6ada req-490ffeb9-196c-4cff-b5fc-131425e0b891 service nova] Acquiring lock "refresh_cache-2afc50fb-138c-4523-b086-db11026ae52f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.060810] env[63293]: DEBUG oslo_concurrency.lockutils [req-1a1f5748-e062-4289-a9c5-5e16f5fb6ada req-490ffeb9-196c-4cff-b5fc-131425e0b891 service nova] Acquired lock "refresh_cache-2afc50fb-138c-4523-b086-db11026ae52f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.060810] env[63293]: DEBUG nova.network.neutron [req-1a1f5748-e062-4289-a9c5-5e16f5fb6ada req-490ffeb9-196c-4cff-b5fc-131425e0b891 service nova] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Refreshing network info cache for port 11eb4a89-6521-4f4d-867e-977a50fd05ab {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 713.420266] env[63293]: DEBUG nova.scheduler.client.report [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.434662] env[63293]: DEBUG nova.network.neutron [-] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.589242] env[63293]: DEBUG nova.network.neutron [req-1a1f5748-e062-4289-a9c5-5e16f5fb6ada req-490ffeb9-196c-4cff-b5fc-131425e0b891 service nova] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.737117] env[63293]: DEBUG nova.network.neutron [req-1a1f5748-e062-4289-a9c5-5e16f5fb6ada req-490ffeb9-196c-4cff-b5fc-131425e0b891 service nova] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.848434] env[63293]: ERROR nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6c767097-c348-4bbf-b8c9-a1d2ab41926d, please check neutron logs for more information. [ 713.848434] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 713.848434] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.848434] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 713.848434] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.848434] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 713.848434] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.848434] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 713.848434] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.848434] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 713.848434] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.848434] env[63293]: ERROR nova.compute.manager raise self.value [ 713.848434] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.848434] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 713.848434] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.848434] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 713.848935] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.848935] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 713.848935] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6c767097-c348-4bbf-b8c9-a1d2ab41926d, please check neutron logs for more information. [ 713.848935] env[63293]: ERROR nova.compute.manager [ 713.848935] env[63293]: Traceback (most recent call last): [ 713.848935] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 713.848935] env[63293]: listener.cb(fileno) [ 713.848935] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.848935] env[63293]: result = function(*args, **kwargs) [ 713.848935] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 713.848935] env[63293]: return func(*args, **kwargs) [ 713.848935] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.848935] env[63293]: raise e [ 713.848935] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.848935] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 713.848935] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.848935] env[63293]: created_port_ids = self._update_ports_for_instance( [ 713.848935] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.848935] env[63293]: with excutils.save_and_reraise_exception(): [ 713.848935] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.848935] env[63293]: self.force_reraise() [ 713.848935] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.848935] env[63293]: raise self.value [ 713.848935] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.848935] env[63293]: updated_port = self._update_port( [ 713.848935] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.848935] env[63293]: _ensure_no_port_binding_failure(port) [ 713.848935] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.848935] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 713.849769] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 6c767097-c348-4bbf-b8c9-a1d2ab41926d, please check neutron logs for more information. [ 713.849769] env[63293]: Removing descriptor: 17 [ 713.925580] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.926430] env[63293]: ERROR nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ca939f80-43b1-4669-9070-f2300e069e4e, please check neutron logs for more information. [ 713.926430] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Traceback (most recent call last): [ 713.926430] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 713.926430] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] self.driver.spawn(context, instance, image_meta, [ 713.926430] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 713.926430] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.926430] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.926430] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] vm_ref = self.build_virtual_machine(instance, [ 713.926430] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.926430] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.926430] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] for vif in network_info: [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] return self._sync_wrapper(fn, *args, **kwargs) [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] self.wait() [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] self[:] = self._gt.wait() [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] return self._exit_event.wait() [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] current.throw(*self._exc) [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.927017] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] result = function(*args, **kwargs) [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] return func(*args, **kwargs) [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] raise e [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] nwinfo = self.network_api.allocate_for_instance( [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] created_port_ids = self._update_ports_for_instance( [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] with excutils.save_and_reraise_exception(): [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] self.force_reraise() [ 713.927601] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.928319] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] raise self.value [ 713.928319] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.928319] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] updated_port = self._update_port( [ 713.928319] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.928319] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] _ensure_no_port_binding_failure(port) [ 713.928319] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.928319] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] raise exception.PortBindingFailed(port_id=port['id']) [ 713.928319] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] nova.exception.PortBindingFailed: Binding failed for port ca939f80-43b1-4669-9070-f2300e069e4e, please check neutron logs for more information. [ 713.928319] env[63293]: ERROR nova.compute.manager [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] [ 713.928319] env[63293]: DEBUG nova.compute.utils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Binding failed for port ca939f80-43b1-4669-9070-f2300e069e4e, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 713.929469] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.739s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.931398] env[63293]: INFO nova.compute.claims [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.934421] env[63293]: DEBUG nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Build of instance 92208bca-bbac-48fa-83db-3f721e6c80c2 was re-scheduled: Binding failed for port ca939f80-43b1-4669-9070-f2300e069e4e, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 713.934874] env[63293]: DEBUG nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 713.935111] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquiring lock "refresh_cache-92208bca-bbac-48fa-83db-3f721e6c80c2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.935262] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquired lock "refresh_cache-92208bca-bbac-48fa-83db-3f721e6c80c2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.935485] env[63293]: DEBUG nova.network.neutron [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.936481] env[63293]: INFO nova.compute.manager [-] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Took 1.04 seconds to deallocate network for instance. [ 713.940095] env[63293]: DEBUG nova.compute.claims [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 713.940266] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.949162] env[63293]: DEBUG nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 713.981346] env[63293]: DEBUG nova.virt.hardware [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 713.981571] env[63293]: DEBUG nova.virt.hardware [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 713.981731] env[63293]: DEBUG nova.virt.hardware [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 713.981909] env[63293]: DEBUG nova.virt.hardware [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 713.982153] env[63293]: DEBUG nova.virt.hardware [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 713.982380] env[63293]: DEBUG nova.virt.hardware [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 713.982601] env[63293]: DEBUG nova.virt.hardware [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 713.982832] env[63293]: DEBUG nova.virt.hardware [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 713.983033] env[63293]: DEBUG nova.virt.hardware [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 713.983206] env[63293]: DEBUG nova.virt.hardware [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 713.983436] env[63293]: DEBUG nova.virt.hardware [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 713.984528] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68619af-538f-4db8-ac25-ca4607b3a717 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.992884] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9ae5e5-f10f-433c-bee8-91146f30ada0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.007564] env[63293]: ERROR nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6c767097-c348-4bbf-b8c9-a1d2ab41926d, please check neutron logs for more information. [ 714.007564] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Traceback (most recent call last): [ 714.007564] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 714.007564] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] yield resources [ 714.007564] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.007564] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] self.driver.spawn(context, instance, image_meta, [ 714.007564] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 714.007564] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.007564] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.007564] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] vm_ref = self.build_virtual_machine(instance, [ 714.007564] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] for vif in network_info: [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] return self._sync_wrapper(fn, *args, **kwargs) [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] self.wait() [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] self[:] = self._gt.wait() [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] return self._exit_event.wait() [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.008175] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] current.throw(*self._exc) [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] result = function(*args, **kwargs) [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] return func(*args, **kwargs) [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] raise e [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] nwinfo = self.network_api.allocate_for_instance( [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] created_port_ids = self._update_ports_for_instance( [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] with excutils.save_and_reraise_exception(): [ 714.008612] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.009201] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] self.force_reraise() [ 714.009201] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.009201] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] raise self.value [ 714.009201] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.009201] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] updated_port = self._update_port( [ 714.009201] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.009201] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] _ensure_no_port_binding_failure(port) [ 714.009201] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.009201] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] raise exception.PortBindingFailed(port_id=port['id']) [ 714.009201] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] nova.exception.PortBindingFailed: Binding failed for port 6c767097-c348-4bbf-b8c9-a1d2ab41926d, please check neutron logs for more information. [ 714.009201] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] [ 714.009201] env[63293]: INFO nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Terminating instance [ 714.009906] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Acquiring lock "refresh_cache-5bf150a5-ad49-4bf8-a8fa-e376e0f00926" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.009977] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Acquired lock "refresh_cache-5bf150a5-ad49-4bf8-a8fa-e376e0f00926" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.010109] env[63293]: DEBUG nova.network.neutron [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 714.240675] env[63293]: DEBUG oslo_concurrency.lockutils [req-1a1f5748-e062-4289-a9c5-5e16f5fb6ada req-490ffeb9-196c-4cff-b5fc-131425e0b891 service nova] Releasing lock "refresh_cache-2afc50fb-138c-4523-b086-db11026ae52f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.240945] env[63293]: DEBUG nova.compute.manager [req-1a1f5748-e062-4289-a9c5-5e16f5fb6ada req-490ffeb9-196c-4cff-b5fc-131425e0b891 service nova] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Received event network-vif-deleted-11eb4a89-6521-4f4d-867e-977a50fd05ab {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.461021] env[63293]: DEBUG nova.network.neutron [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.536839] env[63293]: DEBUG nova.network.neutron [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.572317] env[63293]: DEBUG nova.network.neutron [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.634260] env[63293]: DEBUG nova.network.neutron [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.074625] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Releasing lock "refresh_cache-92208bca-bbac-48fa-83db-3f721e6c80c2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.074923] env[63293]: DEBUG nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 715.074923] env[63293]: DEBUG nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.075411] env[63293]: DEBUG nova.network.neutron [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 715.099918] env[63293]: DEBUG nova.network.neutron [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.127381] env[63293]: DEBUG nova.compute.manager [req-c06ae512-b0a7-4196-bad2-2f5524faef3f req-bf0c6c2b-afd0-4174-92cc-9b83be02dbc4 service nova] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Received event network-changed-6c767097-c348-4bbf-b8c9-a1d2ab41926d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 715.127573] env[63293]: DEBUG nova.compute.manager [req-c06ae512-b0a7-4196-bad2-2f5524faef3f req-bf0c6c2b-afd0-4174-92cc-9b83be02dbc4 service nova] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Refreshing instance network info cache due to event network-changed-6c767097-c348-4bbf-b8c9-a1d2ab41926d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 715.127761] env[63293]: DEBUG oslo_concurrency.lockutils [req-c06ae512-b0a7-4196-bad2-2f5524faef3f req-bf0c6c2b-afd0-4174-92cc-9b83be02dbc4 service nova] Acquiring lock "refresh_cache-5bf150a5-ad49-4bf8-a8fa-e376e0f00926" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.137480] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Releasing lock "refresh_cache-5bf150a5-ad49-4bf8-a8fa-e376e0f00926" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.137865] env[63293]: DEBUG nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 715.138108] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.141064] env[63293]: DEBUG oslo_concurrency.lockutils [req-c06ae512-b0a7-4196-bad2-2f5524faef3f req-bf0c6c2b-afd0-4174-92cc-9b83be02dbc4 service nova] Acquired lock "refresh_cache-5bf150a5-ad49-4bf8-a8fa-e376e0f00926" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.141252] env[63293]: DEBUG nova.network.neutron [req-c06ae512-b0a7-4196-bad2-2f5524faef3f req-bf0c6c2b-afd0-4174-92cc-9b83be02dbc4 service nova] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Refreshing network info cache for port 6c767097-c348-4bbf-b8c9-a1d2ab41926d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 715.142276] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ce2ed9a1-80ab-4e60-a22e-545c4a387bfd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.156094] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba57bb2-6bcd-4ac7-88d5-5984c060e75c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.181404] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5bf150a5-ad49-4bf8-a8fa-e376e0f00926 could not be found. [ 715.181615] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 715.181790] env[63293]: INFO nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Took 0.04 seconds to destroy the instance on the hypervisor. [ 715.182039] env[63293]: DEBUG oslo.service.loopingcall [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.184478] env[63293]: DEBUG nova.compute.manager [-] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.184595] env[63293]: DEBUG nova.network.neutron [-] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 715.203040] env[63293]: DEBUG nova.network.neutron [-] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.339149] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2470894d-d43c-4c22-b8cd-8ee4bbe2a6d8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.347988] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c9b61c-d63a-426c-8dc9-9857debe69d8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.391707] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e30f9b-a347-4106-bc20-479466df5044 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.402875] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222d238d-36f4-41c7-b42f-794f56009b1b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.416654] env[63293]: DEBUG nova.compute.provider_tree [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.487305] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquiring lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.487520] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.607304] env[63293]: DEBUG nova.network.neutron [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.671694] env[63293]: DEBUG nova.network.neutron [req-c06ae512-b0a7-4196-bad2-2f5524faef3f req-bf0c6c2b-afd0-4174-92cc-9b83be02dbc4 service nova] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.705436] env[63293]: DEBUG nova.network.neutron [-] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.763613] env[63293]: DEBUG nova.network.neutron [req-c06ae512-b0a7-4196-bad2-2f5524faef3f req-bf0c6c2b-afd0-4174-92cc-9b83be02dbc4 service nova] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.919668] env[63293]: DEBUG nova.scheduler.client.report [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.110238] env[63293]: INFO nova.compute.manager [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 92208bca-bbac-48fa-83db-3f721e6c80c2] Took 1.03 seconds to deallocate network for instance. [ 716.208253] env[63293]: INFO nova.compute.manager [-] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Took 1.02 seconds to deallocate network for instance. [ 716.210685] env[63293]: DEBUG nova.compute.claims [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 716.210771] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.266857] env[63293]: DEBUG oslo_concurrency.lockutils [req-c06ae512-b0a7-4196-bad2-2f5524faef3f req-bf0c6c2b-afd0-4174-92cc-9b83be02dbc4 service nova] Releasing lock "refresh_cache-5bf150a5-ad49-4bf8-a8fa-e376e0f00926" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.267131] env[63293]: DEBUG nova.compute.manager [req-c06ae512-b0a7-4196-bad2-2f5524faef3f req-bf0c6c2b-afd0-4174-92cc-9b83be02dbc4 service nova] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Received event network-vif-deleted-6c767097-c348-4bbf-b8c9-a1d2ab41926d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.426494] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.427020] env[63293]: DEBUG nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 716.429883] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.228s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.934429] env[63293]: DEBUG nova.compute.utils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.938942] env[63293]: DEBUG nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 716.939106] env[63293]: DEBUG nova.network.neutron [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 716.979377] env[63293]: DEBUG nova.policy [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b94ec4d2961a43ca9a331a6051fa774d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7744835de0bc4e30b1b29ee5a439cd0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.152791] env[63293]: INFO nova.scheduler.client.report [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Deleted allocations for instance 92208bca-bbac-48fa-83db-3f721e6c80c2 [ 717.340445] env[63293]: DEBUG nova.network.neutron [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Successfully created port: b56a46a2-594c-4c94-a964-07103b663412 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.364332] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839332d6-dd6b-4729-aa76-e2b8f24cebf3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.372074] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d985a848-8ea6-4cfa-93f5-227e4d535fcf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.401761] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dea9156-6b38-42be-ba5d-fac1e8369ea2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.409046] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8522a4-cde0-41e9-af88-e5e740103080 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.421890] env[63293]: DEBUG nova.compute.provider_tree [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.439570] env[63293]: DEBUG nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 717.665643] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1f8f87c0-92a6-4257-a901-942250ae339f tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "92208bca-bbac-48fa-83db-3f721e6c80c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.631s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.924839] env[63293]: DEBUG nova.scheduler.client.report [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.169143] env[63293]: DEBUG nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 718.173436] env[63293]: DEBUG nova.compute.manager [req-5530699d-e4a9-425f-96ee-d1927acc1159 req-349fd71e-876f-4174-9f79-d86aef349a6c service nova] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Received event network-changed-b56a46a2-594c-4c94-a964-07103b663412 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.173642] env[63293]: DEBUG nova.compute.manager [req-5530699d-e4a9-425f-96ee-d1927acc1159 req-349fd71e-876f-4174-9f79-d86aef349a6c service nova] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Refreshing instance network info cache due to event network-changed-b56a46a2-594c-4c94-a964-07103b663412. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 718.173831] env[63293]: DEBUG oslo_concurrency.lockutils [req-5530699d-e4a9-425f-96ee-d1927acc1159 req-349fd71e-876f-4174-9f79-d86aef349a6c service nova] Acquiring lock "refresh_cache-e86185ea-7aa6-43ad-920d-645e18194b04" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.173955] env[63293]: DEBUG oslo_concurrency.lockutils [req-5530699d-e4a9-425f-96ee-d1927acc1159 req-349fd71e-876f-4174-9f79-d86aef349a6c service nova] Acquired lock "refresh_cache-e86185ea-7aa6-43ad-920d-645e18194b04" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.174130] env[63293]: DEBUG nova.network.neutron [req-5530699d-e4a9-425f-96ee-d1927acc1159 req-349fd71e-876f-4174-9f79-d86aef349a6c service nova] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Refreshing network info cache for port b56a46a2-594c-4c94-a964-07103b663412 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 718.371280] env[63293]: ERROR nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b56a46a2-594c-4c94-a964-07103b663412, please check neutron logs for more information. [ 718.371280] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.371280] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.371280] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.371280] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.371280] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.371280] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.371280] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.371280] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.371280] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 718.371280] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.371280] env[63293]: ERROR nova.compute.manager raise self.value [ 718.371280] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.371280] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.371280] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.371280] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.371824] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.371824] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.371824] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b56a46a2-594c-4c94-a964-07103b663412, please check neutron logs for more information. [ 718.371824] env[63293]: ERROR nova.compute.manager [ 718.371824] env[63293]: Traceback (most recent call last): [ 718.371824] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.371824] env[63293]: listener.cb(fileno) [ 718.371824] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.371824] env[63293]: result = function(*args, **kwargs) [ 718.371824] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.371824] env[63293]: return func(*args, **kwargs) [ 718.371824] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.371824] env[63293]: raise e [ 718.371824] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.371824] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 718.371824] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.371824] env[63293]: created_port_ids = self._update_ports_for_instance( [ 718.371824] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.371824] env[63293]: with excutils.save_and_reraise_exception(): [ 718.371824] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.371824] env[63293]: self.force_reraise() [ 718.371824] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.371824] env[63293]: raise self.value [ 718.371824] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.371824] env[63293]: updated_port = self._update_port( [ 718.371824] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.371824] env[63293]: _ensure_no_port_binding_failure(port) [ 718.371824] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.371824] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.372820] env[63293]: nova.exception.PortBindingFailed: Binding failed for port b56a46a2-594c-4c94-a964-07103b663412, please check neutron logs for more information. [ 718.372820] env[63293]: Removing descriptor: 15 [ 718.430152] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.430797] env[63293]: ERROR nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 67f48fe2-558f-4c91-8a40-fc8b5a6e0e97, please check neutron logs for more information. [ 718.430797] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Traceback (most recent call last): [ 718.430797] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.430797] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] self.driver.spawn(context, instance, image_meta, [ 718.430797] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 718.430797] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.430797] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.430797] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] vm_ref = self.build_virtual_machine(instance, [ 718.430797] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.430797] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.430797] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] for vif in network_info: [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] return self._sync_wrapper(fn, *args, **kwargs) [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] self.wait() [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] self[:] = self._gt.wait() [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] return self._exit_event.wait() [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] result = hub.switch() [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 718.431204] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] return self.greenlet.switch() [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] result = function(*args, **kwargs) [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] return func(*args, **kwargs) [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] raise e [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] nwinfo = self.network_api.allocate_for_instance( [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] created_port_ids = self._update_ports_for_instance( [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] with excutils.save_and_reraise_exception(): [ 718.431621] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.432046] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] self.force_reraise() [ 718.432046] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.432046] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] raise self.value [ 718.432046] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.432046] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] updated_port = self._update_port( [ 718.432046] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.432046] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] _ensure_no_port_binding_failure(port) [ 718.432046] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.432046] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] raise exception.PortBindingFailed(port_id=port['id']) [ 718.432046] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] nova.exception.PortBindingFailed: Binding failed for port 67f48fe2-558f-4c91-8a40-fc8b5a6e0e97, please check neutron logs for more information. [ 718.432046] env[63293]: ERROR nova.compute.manager [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] [ 718.432406] env[63293]: DEBUG nova.compute.utils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Binding failed for port 67f48fe2-558f-4c91-8a40-fc8b5a6e0e97, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 718.432853] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.633s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.436728] env[63293]: DEBUG nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Build of instance 305e3a9c-27cf-47b8-8825-09e5abe220ee was re-scheduled: Binding failed for port 67f48fe2-558f-4c91-8a40-fc8b5a6e0e97, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 718.436728] env[63293]: DEBUG nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 718.436728] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquiring lock "refresh_cache-305e3a9c-27cf-47b8-8825-09e5abe220ee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.436728] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Acquired lock "refresh_cache-305e3a9c-27cf-47b8-8825-09e5abe220ee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.436976] env[63293]: DEBUG nova.network.neutron [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.448043] env[63293]: DEBUG nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 718.482303] env[63293]: DEBUG nova.virt.hardware [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.482554] env[63293]: DEBUG nova.virt.hardware [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.482709] env[63293]: DEBUG nova.virt.hardware [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.482891] env[63293]: DEBUG nova.virt.hardware [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.483052] env[63293]: DEBUG nova.virt.hardware [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.483206] env[63293]: DEBUG nova.virt.hardware [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.483414] env[63293]: DEBUG nova.virt.hardware [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.483571] env[63293]: DEBUG nova.virt.hardware [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.483733] env[63293]: DEBUG nova.virt.hardware [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.483915] env[63293]: DEBUG nova.virt.hardware [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.484418] env[63293]: DEBUG nova.virt.hardware [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.485448] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8692026-4396-4e21-a159-a01fa9b5523c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.495015] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbaadb2f-1af3-411e-8135-ad6f42bbfe5d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.509498] env[63293]: ERROR nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b56a46a2-594c-4c94-a964-07103b663412, please check neutron logs for more information. [ 718.509498] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Traceback (most recent call last): [ 718.509498] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 718.509498] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] yield resources [ 718.509498] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.509498] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] self.driver.spawn(context, instance, image_meta, [ 718.509498] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 718.509498] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.509498] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.509498] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] vm_ref = self.build_virtual_machine(instance, [ 718.509498] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] for vif in network_info: [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] return self._sync_wrapper(fn, *args, **kwargs) [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] self.wait() [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] self[:] = self._gt.wait() [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] return self._exit_event.wait() [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 718.509898] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] current.throw(*self._exc) [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] result = function(*args, **kwargs) [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] return func(*args, **kwargs) [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] raise e [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] nwinfo = self.network_api.allocate_for_instance( [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] created_port_ids = self._update_ports_for_instance( [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] with excutils.save_and_reraise_exception(): [ 718.510287] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.510709] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] self.force_reraise() [ 718.510709] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.510709] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] raise self.value [ 718.510709] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.510709] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] updated_port = self._update_port( [ 718.510709] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.510709] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] _ensure_no_port_binding_failure(port) [ 718.510709] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.510709] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] raise exception.PortBindingFailed(port_id=port['id']) [ 718.510709] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] nova.exception.PortBindingFailed: Binding failed for port b56a46a2-594c-4c94-a964-07103b663412, please check neutron logs for more information. [ 718.510709] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] [ 718.510709] env[63293]: INFO nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Terminating instance [ 718.511812] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "refresh_cache-e86185ea-7aa6-43ad-920d-645e18194b04" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.695618] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.696456] env[63293]: DEBUG nova.network.neutron [req-5530699d-e4a9-425f-96ee-d1927acc1159 req-349fd71e-876f-4174-9f79-d86aef349a6c service nova] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.786931] env[63293]: DEBUG nova.network.neutron [req-5530699d-e4a9-425f-96ee-d1927acc1159 req-349fd71e-876f-4174-9f79-d86aef349a6c service nova] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.957760] env[63293]: DEBUG nova.network.neutron [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.004150] env[63293]: DEBUG nova.network.neutron [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.281855] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd999fc8-42ef-4e01-bbba-89a1038c8dcf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.289306] env[63293]: DEBUG oslo_concurrency.lockutils [req-5530699d-e4a9-425f-96ee-d1927acc1159 req-349fd71e-876f-4174-9f79-d86aef349a6c service nova] Releasing lock "refresh_cache-e86185ea-7aa6-43ad-920d-645e18194b04" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.289749] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "refresh_cache-e86185ea-7aa6-43ad-920d-645e18194b04" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.289930] env[63293]: DEBUG nova.network.neutron [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.291723] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f4f417-3e07-49c5-8036-5a6c4ce18922 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.324084] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11ca64c-de88-4417-87df-238749aaabe2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.332857] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7c1e93-d6e1-4d89-9383-d0ae1b338f02 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.346533] env[63293]: DEBUG nova.compute.provider_tree [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.507786] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Releasing lock "refresh_cache-305e3a9c-27cf-47b8-8825-09e5abe220ee" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.508031] env[63293]: DEBUG nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 719.508274] env[63293]: DEBUG nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.508474] env[63293]: DEBUG nova.network.neutron [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.528361] env[63293]: DEBUG nova.network.neutron [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.810968] env[63293]: DEBUG nova.network.neutron [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.850787] env[63293]: DEBUG nova.scheduler.client.report [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.883155] env[63293]: DEBUG nova.network.neutron [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.035032] env[63293]: DEBUG nova.network.neutron [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.207391] env[63293]: DEBUG nova.compute.manager [req-80e0a088-bd8f-44fe-96e7-8fa1b89ce10a req-b5458d6d-6bd8-4566-9d7d-43f84e3bbcc7 service nova] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Received event network-vif-deleted-b56a46a2-594c-4c94-a964-07103b663412 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.354768] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.921s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.356824] env[63293]: ERROR nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 95b75e3a-aced-4da6-8f6f-628e8a758fa7, please check neutron logs for more information. [ 720.356824] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Traceback (most recent call last): [ 720.356824] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.356824] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] self.driver.spawn(context, instance, image_meta, [ 720.356824] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 720.356824] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.356824] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.356824] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] vm_ref = self.build_virtual_machine(instance, [ 720.356824] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.356824] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.356824] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] for vif in network_info: [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] return self._sync_wrapper(fn, *args, **kwargs) [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] self.wait() [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] self[:] = self._gt.wait() [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] return self._exit_event.wait() [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] result = hub.switch() [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 720.357287] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] return self.greenlet.switch() [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] result = function(*args, **kwargs) [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] return func(*args, **kwargs) [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] raise e [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] nwinfo = self.network_api.allocate_for_instance( [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] created_port_ids = self._update_ports_for_instance( [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] with excutils.save_and_reraise_exception(): [ 720.357718] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.358194] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] self.force_reraise() [ 720.358194] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.358194] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] raise self.value [ 720.358194] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.358194] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] updated_port = self._update_port( [ 720.358194] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.358194] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] _ensure_no_port_binding_failure(port) [ 720.358194] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.358194] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] raise exception.PortBindingFailed(port_id=port['id']) [ 720.358194] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] nova.exception.PortBindingFailed: Binding failed for port 95b75e3a-aced-4da6-8f6f-628e8a758fa7, please check neutron logs for more information. [ 720.358194] env[63293]: ERROR nova.compute.manager [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] [ 720.358769] env[63293]: DEBUG nova.compute.utils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Binding failed for port 95b75e3a-aced-4da6-8f6f-628e8a758fa7, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 720.358769] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.051s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.360504] env[63293]: DEBUG nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Build of instance a63744c8-b9bf-4a89-9319-061d53de6b5f was re-scheduled: Binding failed for port 95b75e3a-aced-4da6-8f6f-628e8a758fa7, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 720.360944] env[63293]: DEBUG nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 720.361183] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Acquiring lock "refresh_cache-a63744c8-b9bf-4a89-9319-061d53de6b5f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.361333] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Acquired lock "refresh_cache-a63744c8-b9bf-4a89-9319-061d53de6b5f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.362291] env[63293]: DEBUG nova.network.neutron [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.386030] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "refresh_cache-e86185ea-7aa6-43ad-920d-645e18194b04" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.386170] env[63293]: DEBUG nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.386354] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.386924] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bc5863d-cb4c-4b46-886c-839bb0921088 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.397068] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe8f299-79c6-4843-ad67-43858f366bd8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.420993] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e86185ea-7aa6-43ad-920d-645e18194b04 could not be found. [ 720.420993] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 720.421245] env[63293]: INFO nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Took 0.03 seconds to destroy the instance on the hypervisor. [ 720.421431] env[63293]: DEBUG oslo.service.loopingcall [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.421899] env[63293]: DEBUG nova.compute.manager [-] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.421992] env[63293]: DEBUG nova.network.neutron [-] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.437191] env[63293]: DEBUG nova.network.neutron [-] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.537200] env[63293]: INFO nova.compute.manager [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] [instance: 305e3a9c-27cf-47b8-8825-09e5abe220ee] Took 1.03 seconds to deallocate network for instance. [ 720.880281] env[63293]: DEBUG nova.network.neutron [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.940496] env[63293]: DEBUG nova.network.neutron [-] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.956121] env[63293]: DEBUG nova.network.neutron [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.194586] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6506349-f353-4934-bd4c-e01e179a208a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.202693] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59e0aaf-4610-45dd-a612-95d5b80bf9cf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.232026] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e9f20a-1b3b-4885-8afb-74943f37f709 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.239423] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215c7bf8-389d-4ed3-bd38-5d986d293274 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.252226] env[63293]: DEBUG nova.compute.provider_tree [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.444597] env[63293]: INFO nova.compute.manager [-] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Took 1.02 seconds to deallocate network for instance. [ 721.446631] env[63293]: DEBUG nova.compute.claims [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.446725] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.458366] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Releasing lock "refresh_cache-a63744c8-b9bf-4a89-9319-061d53de6b5f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.458577] env[63293]: DEBUG nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 721.458756] env[63293]: DEBUG nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 721.458920] env[63293]: DEBUG nova.network.neutron [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.476069] env[63293]: DEBUG nova.network.neutron [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.563530] env[63293]: INFO nova.scheduler.client.report [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Deleted allocations for instance 305e3a9c-27cf-47b8-8825-09e5abe220ee [ 721.755440] env[63293]: DEBUG nova.scheduler.client.report [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.978895] env[63293]: DEBUG nova.network.neutron [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.071172] env[63293]: DEBUG oslo_concurrency.lockutils [None req-648a3a68-1f88-4a62-a720-fed3ae9429d1 tempest-MigrationsAdminTest-1595084074 tempest-MigrationsAdminTest-1595084074-project-member] Lock "305e3a9c-27cf-47b8-8825-09e5abe220ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.603s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.260137] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.903s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.260785] env[63293]: ERROR nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2117f165-5629-4dbc-8031-aa8d4898becb, please check neutron logs for more information. [ 722.260785] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Traceback (most recent call last): [ 722.260785] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.260785] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] self.driver.spawn(context, instance, image_meta, [ 722.260785] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 722.260785] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.260785] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.260785] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] vm_ref = self.build_virtual_machine(instance, [ 722.260785] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.260785] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.260785] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] for vif in network_info: [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] return self._sync_wrapper(fn, *args, **kwargs) [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] self.wait() [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] self[:] = self._gt.wait() [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] return self._exit_event.wait() [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] current.throw(*self._exc) [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.261154] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] result = function(*args, **kwargs) [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] return func(*args, **kwargs) [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] raise e [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] nwinfo = self.network_api.allocate_for_instance( [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] created_port_ids = self._update_ports_for_instance( [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] with excutils.save_and_reraise_exception(): [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] self.force_reraise() [ 722.261576] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.261987] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] raise self.value [ 722.261987] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.261987] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] updated_port = self._update_port( [ 722.261987] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.261987] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] _ensure_no_port_binding_failure(port) [ 722.261987] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.261987] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] raise exception.PortBindingFailed(port_id=port['id']) [ 722.261987] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] nova.exception.PortBindingFailed: Binding failed for port 2117f165-5629-4dbc-8031-aa8d4898becb, please check neutron logs for more information. [ 722.261987] env[63293]: ERROR nova.compute.manager [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] [ 722.261987] env[63293]: DEBUG nova.compute.utils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Binding failed for port 2117f165-5629-4dbc-8031-aa8d4898becb, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 722.262689] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.515s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.265773] env[63293]: DEBUG nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Build of instance 8b4b041b-5d3f-479a-8241-c75cce650d1e was re-scheduled: Binding failed for port 2117f165-5629-4dbc-8031-aa8d4898becb, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 722.266789] env[63293]: DEBUG nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 722.266789] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Acquiring lock "refresh_cache-8b4b041b-5d3f-479a-8241-c75cce650d1e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.266789] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Acquired lock "refresh_cache-8b4b041b-5d3f-479a-8241-c75cce650d1e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.266789] env[63293]: DEBUG nova.network.neutron [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 722.482942] env[63293]: INFO nova.compute.manager [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] [instance: a63744c8-b9bf-4a89-9319-061d53de6b5f] Took 1.02 seconds to deallocate network for instance. [ 722.576284] env[63293]: DEBUG nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 722.786056] env[63293]: DEBUG nova.network.neutron [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.882740] env[63293]: DEBUG nova.network.neutron [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.096579] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.109616] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14539ee-d639-4b73-814b-3e7326e5767f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.117707] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd81c81-f6a5-4447-a277-1ef6efcea3cc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.148308] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cadeda4-a05f-4d89-9ed5-a21eb06bb818 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.155919] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0150669c-ceca-4187-9e70-aac49fd1ea4d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.169808] env[63293]: DEBUG nova.compute.provider_tree [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.386613] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Releasing lock "refresh_cache-8b4b041b-5d3f-479a-8241-c75cce650d1e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.386856] env[63293]: DEBUG nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 723.387050] env[63293]: DEBUG nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.387229] env[63293]: DEBUG nova.network.neutron [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 723.402040] env[63293]: DEBUG nova.network.neutron [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.507774] env[63293]: INFO nova.scheduler.client.report [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Deleted allocations for instance a63744c8-b9bf-4a89-9319-061d53de6b5f [ 723.674121] env[63293]: DEBUG nova.scheduler.client.report [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.905332] env[63293]: DEBUG nova.network.neutron [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.017027] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cc263c45-c03c-4b91-8617-792f3469a9a1 tempest-ImagesNegativeTestJSON-859406025 tempest-ImagesNegativeTestJSON-859406025-project-member] Lock "a63744c8-b9bf-4a89-9319-061d53de6b5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.409s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.178808] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.916s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.179541] env[63293]: ERROR nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cd2f776b-4862-4992-964a-4c3ced3420eb, please check neutron logs for more information. [ 724.179541] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Traceback (most recent call last): [ 724.179541] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.179541] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] self.driver.spawn(context, instance, image_meta, [ 724.179541] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 724.179541] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.179541] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.179541] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] vm_ref = self.build_virtual_machine(instance, [ 724.179541] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.179541] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.179541] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] for vif in network_info: [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] return self._sync_wrapper(fn, *args, **kwargs) [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] self.wait() [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] self[:] = self._gt.wait() [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] return self._exit_event.wait() [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] current.throw(*self._exc) [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.179968] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] result = function(*args, **kwargs) [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] return func(*args, **kwargs) [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] raise e [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] nwinfo = self.network_api.allocate_for_instance( [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] created_port_ids = self._update_ports_for_instance( [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] with excutils.save_and_reraise_exception(): [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] self.force_reraise() [ 724.180404] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.180830] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] raise self.value [ 724.180830] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.180830] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] updated_port = self._update_port( [ 724.180830] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.180830] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] _ensure_no_port_binding_failure(port) [ 724.180830] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.180830] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] raise exception.PortBindingFailed(port_id=port['id']) [ 724.180830] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] nova.exception.PortBindingFailed: Binding failed for port cd2f776b-4862-4992-964a-4c3ced3420eb, please check neutron logs for more information. [ 724.180830] env[63293]: ERROR nova.compute.manager [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] [ 724.180830] env[63293]: DEBUG nova.compute.utils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Binding failed for port cd2f776b-4862-4992-964a-4c3ced3420eb, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 724.182036] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.978s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.183229] env[63293]: INFO nova.compute.claims [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.185946] env[63293]: DEBUG nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Build of instance 8c2439e1-26ce-45e5-9179-164c0dbbbf76 was re-scheduled: Binding failed for port cd2f776b-4862-4992-964a-4c3ced3420eb, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 724.186389] env[63293]: DEBUG nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 724.186617] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquiring lock "refresh_cache-8c2439e1-26ce-45e5-9179-164c0dbbbf76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.186760] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Acquired lock "refresh_cache-8c2439e1-26ce-45e5-9179-164c0dbbbf76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.186916] env[63293]: DEBUG nova.network.neutron [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.408383] env[63293]: INFO nova.compute.manager [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] [instance: 8b4b041b-5d3f-479a-8241-c75cce650d1e] Took 1.02 seconds to deallocate network for instance. [ 724.520298] env[63293]: DEBUG nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 724.714007] env[63293]: DEBUG nova.network.neutron [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.872875] env[63293]: DEBUG nova.network.neutron [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.043853] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.379052] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Releasing lock "refresh_cache-8c2439e1-26ce-45e5-9179-164c0dbbbf76" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.379052] env[63293]: DEBUG nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 725.379052] env[63293]: DEBUG nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.379052] env[63293]: DEBUG nova.network.neutron [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.402146] env[63293]: DEBUG nova.network.neutron [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.449433] env[63293]: INFO nova.scheduler.client.report [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Deleted allocations for instance 8b4b041b-5d3f-479a-8241-c75cce650d1e [ 725.558924] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637b709d-07c4-4d6c-8926-0c02cb0a7d6f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.567848] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-988287d1-67be-4469-9e25-67e196a30f69 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.597812] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea94131f-b4fc-4ff6-b588-5862f6f0196d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.605417] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436a1a0a-15e8-4022-b30c-5d0ff17ea2f9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.619552] env[63293]: DEBUG nova.compute.provider_tree [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.904769] env[63293]: DEBUG nova.network.neutron [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.960121] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44c2ff9f-57df-4292-a024-1f08493f2d4f tempest-ServerActionsTestJSON-1920038415 tempest-ServerActionsTestJSON-1920038415-project-member] Lock "8b4b041b-5d3f-479a-8241-c75cce650d1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.984s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.124572] env[63293]: DEBUG nova.scheduler.client.report [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.407480] env[63293]: INFO nova.compute.manager [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] [instance: 8c2439e1-26ce-45e5-9179-164c0dbbbf76] Took 1.03 seconds to deallocate network for instance. [ 726.466760] env[63293]: DEBUG nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.633393] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.638036] env[63293]: DEBUG nova.compute.manager [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.639031] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.359s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.001259] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.154186] env[63293]: DEBUG nova.compute.utils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.155783] env[63293]: DEBUG nova.compute.manager [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Not allocating networking since 'none' was specified. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 727.452349] env[63293]: INFO nova.scheduler.client.report [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Deleted allocations for instance 8c2439e1-26ce-45e5-9179-164c0dbbbf76 [ 727.587224] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ad3fa6-2e19-4d49-b94b-38f5f549e213 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.595565] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e1f6a8-ad7d-4c73-9bee-579c9bea79b1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.629709] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88cd0ecf-cd73-4797-8c27-df5e37fb0e0d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.641024] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54388ea9-b1ff-4fb0-9e11-2a1822e2987d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.655753] env[63293]: DEBUG nova.compute.provider_tree [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.657348] env[63293]: DEBUG nova.compute.manager [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.961717] env[63293]: DEBUG oslo_concurrency.lockutils [None req-054faca9-92ba-48b6-9eac-9dbf5258e951 tempest-ServerRescueNegativeTestJSON-971969572 tempest-ServerRescueNegativeTestJSON-971969572-project-member] Lock "8c2439e1-26ce-45e5-9179-164c0dbbbf76" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 160.654s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.163294] env[63293]: DEBUG nova.scheduler.client.report [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.467130] env[63293]: DEBUG nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.672848] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.034s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.673483] env[63293]: ERROR nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 73e90e52-1f97-4b51-8dfa-3588700c63f3, please check neutron logs for more information. [ 728.673483] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Traceback (most recent call last): [ 728.673483] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.673483] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] self.driver.spawn(context, instance, image_meta, [ 728.673483] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 728.673483] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.673483] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.673483] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] vm_ref = self.build_virtual_machine(instance, [ 728.673483] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.673483] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.673483] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] for vif in network_info: [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] return self._sync_wrapper(fn, *args, **kwargs) [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] self.wait() [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] self[:] = self._gt.wait() [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] return self._exit_event.wait() [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] current.throw(*self._exc) [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.674126] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] result = function(*args, **kwargs) [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] return func(*args, **kwargs) [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] raise e [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] nwinfo = self.network_api.allocate_for_instance( [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] created_port_ids = self._update_ports_for_instance( [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] with excutils.save_and_reraise_exception(): [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] self.force_reraise() [ 728.675167] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.675905] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] raise self.value [ 728.675905] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.675905] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] updated_port = self._update_port( [ 728.675905] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.675905] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] _ensure_no_port_binding_failure(port) [ 728.675905] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.675905] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] raise exception.PortBindingFailed(port_id=port['id']) [ 728.675905] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] nova.exception.PortBindingFailed: Binding failed for port 73e90e52-1f97-4b51-8dfa-3588700c63f3, please check neutron logs for more information. [ 728.675905] env[63293]: ERROR nova.compute.manager [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] [ 728.675905] env[63293]: DEBUG nova.compute.utils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Binding failed for port 73e90e52-1f97-4b51-8dfa-3588700c63f3, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 728.679465] env[63293]: DEBUG nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Build of instance e0e08728-b6c3-4610-bc27-348c5594fefc was re-scheduled: Binding failed for port 73e90e52-1f97-4b51-8dfa-3588700c63f3, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 728.679915] env[63293]: DEBUG nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 728.680701] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Acquiring lock "refresh_cache-e0e08728-b6c3-4610-bc27-348c5594fefc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.680701] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Acquired lock "refresh_cache-e0e08728-b6c3-4610-bc27-348c5594fefc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.680701] env[63293]: DEBUG nova.network.neutron [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.685048] env[63293]: DEBUG nova.compute.manager [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 728.685139] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.745s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.712653] env[63293]: DEBUG nova.network.neutron [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.734340] env[63293]: DEBUG nova.virt.hardware [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.734596] env[63293]: DEBUG nova.virt.hardware [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.734749] env[63293]: DEBUG nova.virt.hardware [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.734930] env[63293]: DEBUG nova.virt.hardware [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.736094] env[63293]: DEBUG nova.virt.hardware [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.736466] env[63293]: DEBUG nova.virt.hardware [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.736815] env[63293]: DEBUG nova.virt.hardware [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.737107] env[63293]: DEBUG nova.virt.hardware [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.737437] env[63293]: DEBUG nova.virt.hardware [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.737761] env[63293]: DEBUG nova.virt.hardware [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.741254] env[63293]: DEBUG nova.virt.hardware [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.743040] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba3e6bb-74d2-4d70-b27b-415b53cafd05 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.756713] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbbe89e6-3d6f-4954-b206-d202ecfb4185 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.772064] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 728.778039] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Creating folder: Project (dd87f6f39ce44a03ab1a55f10f6ca7bb). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 728.778039] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ec27969-2dfe-47c4-9425-c85d5e4196bc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.789920] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Created folder: Project (dd87f6f39ce44a03ab1a55f10f6ca7bb) in parent group-v283678. [ 728.790140] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Creating folder: Instances. Parent ref: group-v283695. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 728.790464] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0bf19f26-8fdf-421f-bd5f-baf9074370fa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.797707] env[63293]: DEBUG nova.network.neutron [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.800022] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Created folder: Instances in parent group-v283695. [ 728.800287] env[63293]: DEBUG oslo.service.loopingcall [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.800685] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 728.800889] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b906162-eaca-4814-80bc-a7dbb300973c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.818739] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 728.818739] env[63293]: value = "task-1327476" [ 728.818739] env[63293]: _type = "Task" [ 728.818739] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.827788] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327476, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.989857] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.302042] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Releasing lock "refresh_cache-e0e08728-b6c3-4610-bc27-348c5594fefc" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.302270] env[63293]: DEBUG nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 729.302989] env[63293]: DEBUG nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.302989] env[63293]: DEBUG nova.network.neutron [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.318185] env[63293]: DEBUG nova.network.neutron [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.337653] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327476, 'name': CreateVM_Task, 'duration_secs': 0.343816} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.337859] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 729.338336] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.338537] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.338896] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 729.339205] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bbf3444-c88a-4075-bf29-2eb951ee7f6d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.345498] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for the task: (returnval){ [ 729.345498] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524fcac7-5e29-59c6-6f03-bf659bf5c4da" [ 729.345498] env[63293]: _type = "Task" [ 729.345498] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.354798] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524fcac7-5e29-59c6-6f03-bf659bf5c4da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.582993] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acdc3d08-5272-485f-8f17-082117d0108a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.593827] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a91bdcd-f71d-47e6-a8ca-b0accb44f86e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.631326] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec38d5f9-c20b-4bf5-9465-ef2ca6841fe6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.642034] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cff8b8-a724-4e04-8eb4-aacc733cc6e6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.654602] env[63293]: DEBUG nova.compute.provider_tree [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.831908] env[63293]: DEBUG nova.network.neutron [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.865707] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524fcac7-5e29-59c6-6f03-bf659bf5c4da, 'name': SearchDatastore_Task, 'duration_secs': 0.015534} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.866362] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.866916] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 729.866916] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.867129] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.867186] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 729.867423] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34355173-692f-48ce-bcbb-2c98ca268ac0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.875257] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 729.875699] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 729.876873] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba0a58e7-0841-49f5-a932-e03208396792 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.881974] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for the task: (returnval){ [ 729.881974] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528c4aa9-510a-5ef5-eaaf-d2465e461236" [ 729.881974] env[63293]: _type = "Task" [ 729.881974] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.890236] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528c4aa9-510a-5ef5-eaaf-d2465e461236, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.158140] env[63293]: DEBUG nova.scheduler.client.report [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.301523] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "ad585ebb-2072-45df-b645-94c9fa93576b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.301899] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.334630] env[63293]: INFO nova.compute.manager [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] [instance: e0e08728-b6c3-4610-bc27-348c5594fefc] Took 1.03 seconds to deallocate network for instance. [ 730.395868] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528c4aa9-510a-5ef5-eaaf-d2465e461236, 'name': SearchDatastore_Task, 'duration_secs': 0.008678} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.396752] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98439fdb-41f4-40f9-8133-7d57c2ddc351 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.402148] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for the task: (returnval){ [ 730.402148] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523efab2-44d2-15a9-3180-d79c11d90463" [ 730.402148] env[63293]: _type = "Task" [ 730.402148] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.411691] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523efab2-44d2-15a9-3180-d79c11d90463, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.665065] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.980s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.665719] env[63293]: ERROR nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 11eb4a89-6521-4f4d-867e-977a50fd05ab, please check neutron logs for more information. [ 730.665719] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Traceback (most recent call last): [ 730.665719] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 730.665719] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] self.driver.spawn(context, instance, image_meta, [ 730.665719] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 730.665719] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.665719] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.665719] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] vm_ref = self.build_virtual_machine(instance, [ 730.665719] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.665719] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.665719] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] for vif in network_info: [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] return self._sync_wrapper(fn, *args, **kwargs) [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] self.wait() [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] self[:] = self._gt.wait() [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] return self._exit_event.wait() [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] result = hub.switch() [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 730.666278] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] return self.greenlet.switch() [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] result = function(*args, **kwargs) [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] return func(*args, **kwargs) [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] raise e [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] nwinfo = self.network_api.allocate_for_instance( [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] created_port_ids = self._update_ports_for_instance( [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] with excutils.save_and_reraise_exception(): [ 730.666905] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.667603] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] self.force_reraise() [ 730.667603] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.667603] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] raise self.value [ 730.667603] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.667603] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] updated_port = self._update_port( [ 730.667603] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.667603] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] _ensure_no_port_binding_failure(port) [ 730.667603] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.667603] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] raise exception.PortBindingFailed(port_id=port['id']) [ 730.667603] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] nova.exception.PortBindingFailed: Binding failed for port 11eb4a89-6521-4f4d-867e-977a50fd05ab, please check neutron logs for more information. [ 730.667603] env[63293]: ERROR nova.compute.manager [instance: 2afc50fb-138c-4523-b086-db11026ae52f] [ 730.668275] env[63293]: DEBUG nova.compute.utils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Binding failed for port 11eb4a89-6521-4f4d-867e-977a50fd05ab, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 730.668275] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.457s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.671166] env[63293]: DEBUG nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Build of instance 2afc50fb-138c-4523-b086-db11026ae52f was re-scheduled: Binding failed for port 11eb4a89-6521-4f4d-867e-977a50fd05ab, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 730.671675] env[63293]: DEBUG nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 730.672060] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Acquiring lock "refresh_cache-2afc50fb-138c-4523-b086-db11026ae52f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.672060] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Acquired lock "refresh_cache-2afc50fb-138c-4523-b086-db11026ae52f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.672489] env[63293]: DEBUG nova.network.neutron [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.912876] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523efab2-44d2-15a9-3180-d79c11d90463, 'name': SearchDatastore_Task, 'duration_secs': 0.010259} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.913203] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.913692] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 0aade4da-882a-4342-aefb-07eb306b17ff/0aade4da-882a-4342-aefb-07eb306b17ff.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 730.913692] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d7ff78d-5a6f-42a8-9ca5-a71fe5328e9e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.920506] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for the task: (returnval){ [ 730.920506] env[63293]: value = "task-1327477" [ 730.920506] env[63293]: _type = "Task" [ 730.920506] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.929563] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.208668] env[63293]: DEBUG nova.network.neutron [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.326412] env[63293]: DEBUG nova.network.neutron [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.381793] env[63293]: INFO nova.scheduler.client.report [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Deleted allocations for instance e0e08728-b6c3-4610-bc27-348c5594fefc [ 731.435189] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.625780] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71be214-d2e8-4bd6-ae22-06189a8a3c79 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.633842] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efc4a0a-b6f2-458a-bc29-6088b51dcbb7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.673996] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716fd90d-2e0a-4312-9f29-7d95b142dae9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.682046] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99da9495-b3b5-4f20-a808-5dd8bd9878f9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.696492] env[63293]: DEBUG nova.compute.provider_tree [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.832641] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Releasing lock "refresh_cache-2afc50fb-138c-4523-b086-db11026ae52f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.832870] env[63293]: DEBUG nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 731.833231] env[63293]: DEBUG nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.833494] env[63293]: DEBUG nova.network.neutron [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.852975] env[63293]: DEBUG nova.network.neutron [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.895863] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6a38e28e-a14b-47b1-865a-95e2d0da667c tempest-ServerExternalEventsTest-1001713914 tempest-ServerExternalEventsTest-1001713914-project-member] Lock "e0e08728-b6c3-4610-bc27-348c5594fefc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.838s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.936568] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.199811] env[63293]: DEBUG nova.scheduler.client.report [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.356689] env[63293]: DEBUG nova.network.neutron [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.401070] env[63293]: DEBUG nova.compute.manager [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 732.432688] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.705457] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.038s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.706117] env[63293]: ERROR nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6c767097-c348-4bbf-b8c9-a1d2ab41926d, please check neutron logs for more information. [ 732.706117] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Traceback (most recent call last): [ 732.706117] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.706117] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] self.driver.spawn(context, instance, image_meta, [ 732.706117] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 732.706117] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.706117] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.706117] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] vm_ref = self.build_virtual_machine(instance, [ 732.706117] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.706117] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.706117] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] for vif in network_info: [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] return self._sync_wrapper(fn, *args, **kwargs) [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] self.wait() [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] self[:] = self._gt.wait() [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] return self._exit_event.wait() [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] current.throw(*self._exc) [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.706735] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] result = function(*args, **kwargs) [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] return func(*args, **kwargs) [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] raise e [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] nwinfo = self.network_api.allocate_for_instance( [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] created_port_ids = self._update_ports_for_instance( [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] with excutils.save_and_reraise_exception(): [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] self.force_reraise() [ 732.707516] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.708150] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] raise self.value [ 732.708150] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.708150] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] updated_port = self._update_port( [ 732.708150] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.708150] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] _ensure_no_port_binding_failure(port) [ 732.708150] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.708150] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] raise exception.PortBindingFailed(port_id=port['id']) [ 732.708150] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] nova.exception.PortBindingFailed: Binding failed for port 6c767097-c348-4bbf-b8c9-a1d2ab41926d, please check neutron logs for more information. [ 732.708150] env[63293]: ERROR nova.compute.manager [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] [ 732.708150] env[63293]: DEBUG nova.compute.utils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Binding failed for port 6c767097-c348-4bbf-b8c9-a1d2ab41926d, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 732.711542] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.013s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.711542] env[63293]: INFO nova.compute.claims [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.714142] env[63293]: DEBUG nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Build of instance 5bf150a5-ad49-4bf8-a8fa-e376e0f00926 was re-scheduled: Binding failed for port 6c767097-c348-4bbf-b8c9-a1d2ab41926d, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 732.714616] env[63293]: DEBUG nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 732.715183] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Acquiring lock "refresh_cache-5bf150a5-ad49-4bf8-a8fa-e376e0f00926" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.715371] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Acquired lock "refresh_cache-5bf150a5-ad49-4bf8-a8fa-e376e0f00926" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.715535] env[63293]: DEBUG nova.network.neutron [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.859549] env[63293]: INFO nova.compute.manager [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] [instance: 2afc50fb-138c-4523-b086-db11026ae52f] Took 1.03 seconds to deallocate network for instance. [ 732.920999] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.934260] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.238952] env[63293]: DEBUG nova.network.neutron [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.337115] env[63293]: DEBUG nova.network.neutron [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.433808] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.690429] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquiring lock "831d7d9b-364f-4c29-bda5-e6c38291a973" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.690562] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lock "831d7d9b-364f-4c29-bda5-e6c38291a973" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.839331] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Releasing lock "refresh_cache-5bf150a5-ad49-4bf8-a8fa-e376e0f00926" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.841830] env[63293]: DEBUG nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 733.841830] env[63293]: DEBUG nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.841830] env[63293]: DEBUG nova.network.neutron [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.860920] env[63293]: DEBUG nova.network.neutron [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.903605] env[63293]: INFO nova.scheduler.client.report [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Deleted allocations for instance 2afc50fb-138c-4523-b086-db11026ae52f [ 733.942616] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.190692] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43bf345-04ec-4b37-9d7a-862a915899f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.201340] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e226cdce-139e-4b05-8483-7cae9e2a155b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.242873] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97bcf10c-c636-4581-b409-364f5bcb5e46 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.250629] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e15c40-9589-487b-b287-bdf6ca3d8630 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.264703] env[63293]: DEBUG nova.compute.provider_tree [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.364466] env[63293]: DEBUG nova.network.neutron [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.412024] env[63293]: DEBUG oslo_concurrency.lockutils [None req-10613ee2-d56c-4db5-8b2a-ac5abccb4c80 tempest-ServersTestFqdnHostnames-977620452 tempest-ServersTestFqdnHostnames-977620452-project-member] Lock "2afc50fb-138c-4523-b086-db11026ae52f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 164.871s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.439577] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.668353] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquiring lock "1e01a75d-122d-4122-9fed-4164d64d4ee8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.668593] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Lock "1e01a75d-122d-4122-9fed-4164d64d4ee8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.771018] env[63293]: DEBUG nova.scheduler.client.report [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.872861] env[63293]: INFO nova.compute.manager [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] [instance: 5bf150a5-ad49-4bf8-a8fa-e376e0f00926] Took 1.03 seconds to deallocate network for instance. [ 734.914576] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.939865] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.274940] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.275619] env[63293]: DEBUG nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 735.278038] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.831s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.447653] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.447653] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.783401] env[63293]: DEBUG nova.compute.utils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.787699] env[63293]: DEBUG nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 735.787877] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 735.894699] env[63293]: DEBUG nova.policy [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3abe71ab8b024338924c7f6389441a93', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d72964783a4d0e95daa2305c6148cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 735.914020] env[63293]: INFO nova.scheduler.client.report [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Deleted allocations for instance 5bf150a5-ad49-4bf8-a8fa-e376e0f00926 [ 735.947396] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.234491] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b24914c-3504-4902-b93c-f333c8b46511 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.242909] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce834c1-cf64-48b9-b93f-65cce5ff718b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.275468] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f94c73-a070-49d9-b30d-38b92377c288 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.283286] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88064357-c7a2-4297-aa68-eb678ed443a8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.288368] env[63293]: DEBUG nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.300425] env[63293]: DEBUG nova.compute.provider_tree [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.424796] env[63293]: DEBUG oslo_concurrency.lockutils [None req-239c47c0-5b80-4824-ab6c-b4c6ea782535 tempest-FloatingIPsAssociationTestJSON-587224058 tempest-FloatingIPsAssociationTestJSON-587224058-project-member] Lock "5bf150a5-ad49-4bf8-a8fa-e376e0f00926" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 165.524s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.442154] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327477, 'name': CopyVirtualDisk_Task, 'duration_secs': 5.261541} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.442612] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 0aade4da-882a-4342-aefb-07eb306b17ff/0aade4da-882a-4342-aefb-07eb306b17ff.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 736.442877] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 736.443681] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2839282-334b-46b4-ac03-2e6b8bc7d968 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.450619] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for the task: (returnval){ [ 736.450619] env[63293]: value = "task-1327478" [ 736.450619] env[63293]: _type = "Task" [ 736.450619] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.458613] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327478, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.610740] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Successfully created port: a0334b0c-e404-47dd-a833-3dcee4aeb4ba {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.808392] env[63293]: DEBUG nova.scheduler.client.report [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.930613] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 736.963283] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327478, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06377} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.963742] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 736.964535] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809c6d6e-ebe4-46f3-8cb9-6a5aa190d311 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.993660] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 0aade4da-882a-4342-aefb-07eb306b17ff/0aade4da-882a-4342-aefb-07eb306b17ff.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 736.994191] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c881a2f-4607-4a2c-b724-6166a97a4477 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.015903] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for the task: (returnval){ [ 737.015903] env[63293]: value = "task-1327479" [ 737.015903] env[63293]: _type = "Task" [ 737.015903] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.025229] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327479, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.217145] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Successfully created port: 0d30e01b-7076-49c1-a5f3-4d08ad24d58a {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.314307] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.036s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.315173] env[63293]: ERROR nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b56a46a2-594c-4c94-a964-07103b663412, please check neutron logs for more information. [ 737.315173] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Traceback (most recent call last): [ 737.315173] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.315173] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] self.driver.spawn(context, instance, image_meta, [ 737.315173] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 737.315173] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.315173] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.315173] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] vm_ref = self.build_virtual_machine(instance, [ 737.315173] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.315173] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.315173] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] for vif in network_info: [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] return self._sync_wrapper(fn, *args, **kwargs) [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] self.wait() [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] self[:] = self._gt.wait() [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] return self._exit_event.wait() [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] current.throw(*self._exc) [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.315614] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] result = function(*args, **kwargs) [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] return func(*args, **kwargs) [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] raise e [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] nwinfo = self.network_api.allocate_for_instance( [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] created_port_ids = self._update_ports_for_instance( [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] with excutils.save_and_reraise_exception(): [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] self.force_reraise() [ 737.316110] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.317792] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] raise self.value [ 737.317792] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.317792] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] updated_port = self._update_port( [ 737.317792] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.317792] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] _ensure_no_port_binding_failure(port) [ 737.317792] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.317792] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] raise exception.PortBindingFailed(port_id=port['id']) [ 737.317792] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] nova.exception.PortBindingFailed: Binding failed for port b56a46a2-594c-4c94-a964-07103b663412, please check neutron logs for more information. [ 737.317792] env[63293]: ERROR nova.compute.manager [instance: e86185ea-7aa6-43ad-920d-645e18194b04] [ 737.317792] env[63293]: DEBUG nova.compute.utils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Binding failed for port b56a46a2-594c-4c94-a964-07103b663412, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 737.319719] env[63293]: DEBUG nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 737.323252] env[63293]: DEBUG nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Build of instance e86185ea-7aa6-43ad-920d-645e18194b04 was re-scheduled: Binding failed for port b56a46a2-594c-4c94-a964-07103b663412, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 737.323843] env[63293]: DEBUG nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 737.323843] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "refresh_cache-e86185ea-7aa6-43ad-920d-645e18194b04" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.323843] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "refresh_cache-e86185ea-7aa6-43ad-920d-645e18194b04" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.323843] env[63293]: DEBUG nova.network.neutron [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.330352] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.233s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.330352] env[63293]: INFO nova.compute.claims [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.353953] env[63293]: DEBUG nova.virt.hardware [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.353953] env[63293]: DEBUG nova.virt.hardware [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.353953] env[63293]: DEBUG nova.virt.hardware [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.353953] env[63293]: DEBUG nova.virt.hardware [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.355403] env[63293]: DEBUG nova.virt.hardware [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.355403] env[63293]: DEBUG nova.virt.hardware [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.355403] env[63293]: DEBUG nova.virt.hardware [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.355403] env[63293]: DEBUG nova.virt.hardware [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.355403] env[63293]: DEBUG nova.virt.hardware [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.355674] env[63293]: DEBUG nova.virt.hardware [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.355674] env[63293]: DEBUG nova.virt.hardware [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.355674] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184220a5-ac79-4ec2-a5a5-1a86e37ac6e8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.363859] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d406a72-408d-463a-90e1-21702db85f6a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.456676] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.534459] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327479, 'name': ReconfigVM_Task, 'duration_secs': 0.331661} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.534854] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 0aade4da-882a-4342-aefb-07eb306b17ff/0aade4da-882a-4342-aefb-07eb306b17ff.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 737.535498] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5a09b06-f8f9-4dbd-9763-07c823632264 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.543021] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for the task: (returnval){ [ 737.543021] env[63293]: value = "task-1327480" [ 737.543021] env[63293]: _type = "Task" [ 737.543021] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.552946] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327480, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.774472] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Successfully created port: e9dfc5a8-cd32-4bd7-b39f-2adbc2febef0 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.867575] env[63293]: DEBUG nova.network.neutron [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.021457] env[63293]: DEBUG nova.network.neutron [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.053308] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327480, 'name': Rename_Task, 'duration_secs': 0.124183} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.053435] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 738.054115] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3fb7db69-e12f-440f-bced-39dd9b209414 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.060312] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for the task: (returnval){ [ 738.060312] env[63293]: value = "task-1327481" [ 738.060312] env[63293]: _type = "Task" [ 738.060312] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.069154] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327481, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.524754] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "refresh_cache-e86185ea-7aa6-43ad-920d-645e18194b04" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.529237] env[63293]: DEBUG nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 738.529650] env[63293]: DEBUG nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.529951] env[63293]: DEBUG nova.network.neutron [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.561159] env[63293]: DEBUG nova.network.neutron [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.572764] env[63293]: DEBUG oslo_vmware.api [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327481, 'name': PowerOnVM_Task, 'duration_secs': 0.482774} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.576611] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 738.576611] env[63293]: INFO nova.compute.manager [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Took 9.89 seconds to spawn the instance on the hypervisor. [ 738.576611] env[63293]: DEBUG nova.compute.manager [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 738.580060] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fc4873-c4ca-4026-818d-d778d8a7d348 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.837098] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf595230-2360-4f12-a8d4-ffb5148be8d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.846973] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e863b1fa-954c-40e2-ba35-5601a9b7289b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.882785] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308587df-e80d-43ed-82e8-3743723e9f0d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.890844] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-511fbc55-68d6-4109-81c6-869013e0ddb2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.908854] env[63293]: DEBUG nova.compute.provider_tree [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.023582] env[63293]: DEBUG nova.compute.manager [req-eb312029-4074-4432-a259-6f1e770ce521 req-4b050b5b-015a-4cc9-bc98-a0337082e3d6 service nova] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Received event network-changed-a0334b0c-e404-47dd-a833-3dcee4aeb4ba {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.023767] env[63293]: DEBUG nova.compute.manager [req-eb312029-4074-4432-a259-6f1e770ce521 req-4b050b5b-015a-4cc9-bc98-a0337082e3d6 service nova] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Refreshing instance network info cache due to event network-changed-a0334b0c-e404-47dd-a833-3dcee4aeb4ba. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 739.023973] env[63293]: DEBUG oslo_concurrency.lockutils [req-eb312029-4074-4432-a259-6f1e770ce521 req-4b050b5b-015a-4cc9-bc98-a0337082e3d6 service nova] Acquiring lock "refresh_cache-db04bc40-1106-46dd-ad9a-f9b11a6c8774" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.024121] env[63293]: DEBUG oslo_concurrency.lockutils [req-eb312029-4074-4432-a259-6f1e770ce521 req-4b050b5b-015a-4cc9-bc98-a0337082e3d6 service nova] Acquired lock "refresh_cache-db04bc40-1106-46dd-ad9a-f9b11a6c8774" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.024277] env[63293]: DEBUG nova.network.neutron [req-eb312029-4074-4432-a259-6f1e770ce521 req-4b050b5b-015a-4cc9-bc98-a0337082e3d6 service nova] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Refreshing network info cache for port a0334b0c-e404-47dd-a833-3dcee4aeb4ba {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 739.065102] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquiring lock "4593cabd-1331-4ee1-99f5-46edeac56859" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.065328] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Lock "4593cabd-1331-4ee1-99f5-46edeac56859" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.068687] env[63293]: DEBUG nova.network.neutron [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.101883] env[63293]: INFO nova.compute.manager [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Took 29.91 seconds to build instance. [ 739.135481] env[63293]: ERROR nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a0334b0c-e404-47dd-a833-3dcee4aeb4ba, please check neutron logs for more information. [ 739.135481] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 739.135481] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.135481] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 739.135481] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.135481] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 739.135481] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.135481] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 739.135481] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.135481] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 739.135481] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.135481] env[63293]: ERROR nova.compute.manager raise self.value [ 739.135481] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.135481] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 739.135481] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.135481] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 739.136295] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.136295] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 739.136295] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a0334b0c-e404-47dd-a833-3dcee4aeb4ba, please check neutron logs for more information. [ 739.136295] env[63293]: ERROR nova.compute.manager [ 739.136295] env[63293]: Traceback (most recent call last): [ 739.136295] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 739.136295] env[63293]: listener.cb(fileno) [ 739.136295] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.136295] env[63293]: result = function(*args, **kwargs) [ 739.136295] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.136295] env[63293]: return func(*args, **kwargs) [ 739.136295] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.136295] env[63293]: raise e [ 739.136295] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.136295] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 739.136295] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.136295] env[63293]: created_port_ids = self._update_ports_for_instance( [ 739.136295] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.136295] env[63293]: with excutils.save_and_reraise_exception(): [ 739.136295] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.136295] env[63293]: self.force_reraise() [ 739.136295] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.136295] env[63293]: raise self.value [ 739.136295] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.136295] env[63293]: updated_port = self._update_port( [ 739.136295] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.136295] env[63293]: _ensure_no_port_binding_failure(port) [ 739.136295] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.136295] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 739.138366] env[63293]: nova.exception.PortBindingFailed: Binding failed for port a0334b0c-e404-47dd-a833-3dcee4aeb4ba, please check neutron logs for more information. [ 739.138366] env[63293]: Removing descriptor: 15 [ 739.138366] env[63293]: ERROR nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a0334b0c-e404-47dd-a833-3dcee4aeb4ba, please check neutron logs for more information. [ 739.138366] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Traceback (most recent call last): [ 739.138366] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 739.138366] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] yield resources [ 739.138366] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.138366] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] self.driver.spawn(context, instance, image_meta, [ 739.138366] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 739.138366] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.138366] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.138366] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] vm_ref = self.build_virtual_machine(instance, [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] for vif in network_info: [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] return self._sync_wrapper(fn, *args, **kwargs) [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] self.wait() [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] self[:] = self._gt.wait() [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] return self._exit_event.wait() [ 739.138753] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] result = hub.switch() [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] return self.greenlet.switch() [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] result = function(*args, **kwargs) [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] return func(*args, **kwargs) [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] raise e [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] nwinfo = self.network_api.allocate_for_instance( [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.139161] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] created_port_ids = self._update_ports_for_instance( [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] with excutils.save_and_reraise_exception(): [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] self.force_reraise() [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] raise self.value [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] updated_port = self._update_port( [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] _ensure_no_port_binding_failure(port) [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.139592] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] raise exception.PortBindingFailed(port_id=port['id']) [ 739.140907] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] nova.exception.PortBindingFailed: Binding failed for port a0334b0c-e404-47dd-a833-3dcee4aeb4ba, please check neutron logs for more information. [ 739.140907] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] [ 739.140907] env[63293]: INFO nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Terminating instance [ 739.140907] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "refresh_cache-db04bc40-1106-46dd-ad9a-f9b11a6c8774" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.411838] env[63293]: DEBUG nova.scheduler.client.report [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.573979] env[63293]: INFO nova.compute.manager [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: e86185ea-7aa6-43ad-920d-645e18194b04] Took 1.04 seconds to deallocate network for instance. [ 739.592225] env[63293]: DEBUG nova.network.neutron [req-eb312029-4074-4432-a259-6f1e770ce521 req-4b050b5b-015a-4cc9-bc98-a0337082e3d6 service nova] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.607495] env[63293]: DEBUG oslo_concurrency.lockutils [None req-eaef8c0e-bd91-4b54-9f7c-e7b325223778 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lock "0aade4da-882a-4342-aefb-07eb306b17ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 162.814s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.911495] env[63293]: DEBUG nova.network.neutron [req-eb312029-4074-4432-a259-6f1e770ce521 req-4b050b5b-015a-4cc9-bc98-a0337082e3d6 service nova] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.918492] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.919209] env[63293]: DEBUG nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 739.921887] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.878s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.923532] env[63293]: INFO nova.compute.claims [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.111196] env[63293]: DEBUG nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 740.416541] env[63293]: DEBUG oslo_concurrency.lockutils [req-eb312029-4074-4432-a259-6f1e770ce521 req-4b050b5b-015a-4cc9-bc98-a0337082e3d6 service nova] Releasing lock "refresh_cache-db04bc40-1106-46dd-ad9a-f9b11a6c8774" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.416948] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquired lock "refresh_cache-db04bc40-1106-46dd-ad9a-f9b11a6c8774" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.417247] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.429812] env[63293]: DEBUG nova.compute.utils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 740.435475] env[63293]: DEBUG nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 740.435475] env[63293]: DEBUG nova.network.neutron [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 740.509395] env[63293]: DEBUG nova.policy [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '99317b99397c403eab3b3571bf844377', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f43f59c7120e45928cb1f3ad16a84b5f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 740.622851] env[63293]: INFO nova.scheduler.client.report [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted allocations for instance e86185ea-7aa6-43ad-920d-645e18194b04 [ 740.653596] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.827935] env[63293]: DEBUG nova.compute.manager [None req-2cf145e5-8309-4896-95d6-c4890597ee01 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 740.829054] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9873482-3b12-493f-8b88-541cb47a28aa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.933270] env[63293]: DEBUG nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 740.980370] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.082240] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquiring lock "0aade4da-882a-4342-aefb-07eb306b17ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.082240] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lock "0aade4da-882a-4342-aefb-07eb306b17ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.082240] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquiring lock "0aade4da-882a-4342-aefb-07eb306b17ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.082240] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lock "0aade4da-882a-4342-aefb-07eb306b17ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.082503] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lock "0aade4da-882a-4342-aefb-07eb306b17ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.087861] env[63293]: INFO nova.compute.manager [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Terminating instance [ 741.090073] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquiring lock "refresh_cache-0aade4da-882a-4342-aefb-07eb306b17ff" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.090073] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquired lock "refresh_cache-0aade4da-882a-4342-aefb-07eb306b17ff" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.090073] env[63293]: DEBUG nova.network.neutron [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 741.105091] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.139406] env[63293]: DEBUG oslo_concurrency.lockutils [None req-727ede00-00b6-4faa-a17f-17bc4aa70b80 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "e86185ea-7aa6-43ad-920d-645e18194b04" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 165.280s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.171510] env[63293]: DEBUG nova.compute.manager [req-aa1aaedb-e89e-4d15-b1a1-ee641e40ba0e req-96533a33-c5f3-4fc7-bd78-7d1f6204b029 service nova] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Received event network-vif-deleted-a0334b0c-e404-47dd-a833-3dcee4aeb4ba {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.345720] env[63293]: INFO nova.compute.manager [None req-2cf145e5-8309-4896-95d6-c4890597ee01 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] instance snapshotting [ 741.346280] env[63293]: DEBUG nova.objects.instance [None req-2cf145e5-8309-4896-95d6-c4890597ee01 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lazy-loading 'flavor' on Instance uuid 0aade4da-882a-4342-aefb-07eb306b17ff {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 741.359552] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfcc8d5-07aa-4d62-94a1-6c8823769c91 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.365686] env[63293]: DEBUG nova.network.neutron [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Successfully created port: 37e8a179-ff61-48b6-8d49-44c09262ea80 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 741.375334] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2bd20b-36d6-46b1-a209-8c30f3d50137 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.414842] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08c3bc4-5de1-4b40-b3b3-0a3b1c93d133 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.422597] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ee0db9-3761-49b0-ac02-67148e30ee56 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.437110] env[63293]: DEBUG nova.compute.provider_tree [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.615080] env[63293]: DEBUG nova.network.neutron [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.615080] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Releasing lock "refresh_cache-db04bc40-1106-46dd-ad9a-f9b11a6c8774" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.615080] env[63293]: DEBUG nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 741.615080] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 741.615080] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d675964-2a89-48de-8fc0-36e7e9b35a5f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.624178] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d028f4-954f-4b58-8559-3703008ae8bd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.645746] env[63293]: DEBUG nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 741.648662] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance db04bc40-1106-46dd-ad9a-f9b11a6c8774 could not be found. [ 741.648861] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 741.649318] env[63293]: INFO nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Took 0.03 seconds to destroy the instance on the hypervisor. [ 741.649318] env[63293]: DEBUG oslo.service.loopingcall [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.650012] env[63293]: DEBUG nova.compute.manager [-] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.650012] env[63293]: DEBUG nova.network.neutron [-] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.709370] env[63293]: DEBUG nova.network.neutron [-] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.747887] env[63293]: DEBUG nova.network.neutron [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.854807] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf5d07c-10d3-4091-9c89-043dba13d769 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.873663] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96313bbd-368e-4d4f-8c7b-fba487de5ffe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.940597] env[63293]: DEBUG nova.scheduler.client.report [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.946615] env[63293]: DEBUG nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 741.980620] env[63293]: DEBUG nova.virt.hardware [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 741.980835] env[63293]: DEBUG nova.virt.hardware [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 741.980994] env[63293]: DEBUG nova.virt.hardware [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.981213] env[63293]: DEBUG nova.virt.hardware [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 741.981399] env[63293]: DEBUG nova.virt.hardware [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.981593] env[63293]: DEBUG nova.virt.hardware [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 741.981856] env[63293]: DEBUG nova.virt.hardware [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 741.982052] env[63293]: DEBUG nova.virt.hardware [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 741.982262] env[63293]: DEBUG nova.virt.hardware [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 741.982996] env[63293]: DEBUG nova.virt.hardware [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 741.982996] env[63293]: DEBUG nova.virt.hardware [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.983753] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2dda510-4d7f-4378-b940-d964d259089d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.992451] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e51de766-1f68-4c78-ac4f-63ebd3dfd153 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.180834] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.252364] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Releasing lock "refresh_cache-0aade4da-882a-4342-aefb-07eb306b17ff" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.252847] env[63293]: DEBUG nova.compute.manager [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 742.253088] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 742.254567] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0e9b78-dbda-456b-9e3c-7d6d8968d786 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.266020] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 742.266020] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00d0a5b0-a0e5-474a-ad9b-07cae73a3c66 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.272150] env[63293]: DEBUG oslo_vmware.api [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for the task: (returnval){ [ 742.272150] env[63293]: value = "task-1327482" [ 742.272150] env[63293]: _type = "Task" [ 742.272150] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.285623] env[63293]: DEBUG oslo_vmware.api [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327482, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.385539] env[63293]: DEBUG nova.compute.manager [None req-2cf145e5-8309-4896-95d6-c4890597ee01 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Instance disappeared during snapshot {{(pid=63293) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 742.446447] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.447736] env[63293]: DEBUG nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 742.452822] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.452s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.454815] env[63293]: INFO nova.compute.claims [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.556823] env[63293]: DEBUG nova.compute.manager [None req-2cf145e5-8309-4896-95d6-c4890597ee01 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Found 0 images (rotation: 2) {{(pid=63293) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 742.783321] env[63293]: DEBUG oslo_vmware.api [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327482, 'name': PowerOffVM_Task, 'duration_secs': 0.200817} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.784305] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 742.784305] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 742.784673] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bbffb964-5bf0-4b62-b564-bdc75b1c97a4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.811018] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 742.811018] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 742.811018] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Deleting the datastore file [datastore1] 0aade4da-882a-4342-aefb-07eb306b17ff {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 742.811018] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a96f0c37-13b6-491a-8265-f3026272841e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.824561] env[63293]: DEBUG oslo_vmware.api [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for the task: (returnval){ [ 742.824561] env[63293]: value = "task-1327484" [ 742.824561] env[63293]: _type = "Task" [ 742.824561] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.832627] env[63293]: DEBUG oslo_vmware.api [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327484, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.902684] env[63293]: DEBUG nova.network.neutron [-] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.964690] env[63293]: DEBUG nova.compute.utils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 742.970995] env[63293]: DEBUG nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 742.970995] env[63293]: DEBUG nova.network.neutron [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 743.057843] env[63293]: DEBUG nova.policy [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd2a9495ab0b4768a4075a71f25aa9a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff8479fc251544e9b4618deccf992754', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 743.275974] env[63293]: ERROR nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 37e8a179-ff61-48b6-8d49-44c09262ea80, please check neutron logs for more information. [ 743.275974] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 743.275974] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.275974] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 743.275974] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.275974] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 743.275974] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.275974] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 743.275974] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.275974] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 743.275974] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.275974] env[63293]: ERROR nova.compute.manager raise self.value [ 743.275974] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.275974] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 743.275974] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.275974] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 743.276600] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.276600] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 743.276600] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 37e8a179-ff61-48b6-8d49-44c09262ea80, please check neutron logs for more information. [ 743.276600] env[63293]: ERROR nova.compute.manager [ 743.276600] env[63293]: Traceback (most recent call last): [ 743.276600] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 743.276600] env[63293]: listener.cb(fileno) [ 743.276600] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.276600] env[63293]: result = function(*args, **kwargs) [ 743.276600] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.276600] env[63293]: return func(*args, **kwargs) [ 743.276600] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.276600] env[63293]: raise e [ 743.276600] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.276600] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 743.276600] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.276600] env[63293]: created_port_ids = self._update_ports_for_instance( [ 743.276600] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.276600] env[63293]: with excutils.save_and_reraise_exception(): [ 743.276600] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.276600] env[63293]: self.force_reraise() [ 743.276600] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.276600] env[63293]: raise self.value [ 743.276600] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.276600] env[63293]: updated_port = self._update_port( [ 743.276600] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.276600] env[63293]: _ensure_no_port_binding_failure(port) [ 743.276600] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.276600] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 743.277630] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 37e8a179-ff61-48b6-8d49-44c09262ea80, please check neutron logs for more information. [ 743.277630] env[63293]: Removing descriptor: 15 [ 743.277630] env[63293]: ERROR nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 37e8a179-ff61-48b6-8d49-44c09262ea80, please check neutron logs for more information. [ 743.277630] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Traceback (most recent call last): [ 743.277630] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 743.277630] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] yield resources [ 743.277630] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.277630] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] self.driver.spawn(context, instance, image_meta, [ 743.277630] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 743.277630] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.277630] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.277630] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] vm_ref = self.build_virtual_machine(instance, [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] for vif in network_info: [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] return self._sync_wrapper(fn, *args, **kwargs) [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] self.wait() [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] self[:] = self._gt.wait() [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] return self._exit_event.wait() [ 743.278045] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] result = hub.switch() [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] return self.greenlet.switch() [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] result = function(*args, **kwargs) [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] return func(*args, **kwargs) [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] raise e [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] nwinfo = self.network_api.allocate_for_instance( [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.278539] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] created_port_ids = self._update_ports_for_instance( [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] with excutils.save_and_reraise_exception(): [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] self.force_reraise() [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] raise self.value [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] updated_port = self._update_port( [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] _ensure_no_port_binding_failure(port) [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.278962] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] raise exception.PortBindingFailed(port_id=port['id']) [ 743.279393] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] nova.exception.PortBindingFailed: Binding failed for port 37e8a179-ff61-48b6-8d49-44c09262ea80, please check neutron logs for more information. [ 743.279393] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] [ 743.279393] env[63293]: INFO nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Terminating instance [ 743.280406] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Acquiring lock "refresh_cache-263d1329-a81a-4d50-91b6-3927bffe8d4c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.280670] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Acquired lock "refresh_cache-263d1329-a81a-4d50-91b6-3927bffe8d4c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.280853] env[63293]: DEBUG nova.network.neutron [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 743.335237] env[63293]: DEBUG oslo_vmware.api [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Task: {'id': task-1327484, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.107444} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.335521] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 743.335712] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 743.335977] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 743.336223] env[63293]: INFO nova.compute.manager [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Took 1.08 seconds to destroy the instance on the hypervisor. [ 743.336561] env[63293]: DEBUG oslo.service.loopingcall [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.336757] env[63293]: DEBUG nova.compute.manager [-] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.336849] env[63293]: DEBUG nova.network.neutron [-] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.353389] env[63293]: DEBUG nova.network.neutron [-] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.404215] env[63293]: INFO nova.compute.manager [-] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Took 1.75 seconds to deallocate network for instance. [ 743.408413] env[63293]: DEBUG nova.compute.claims [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 743.408413] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.471073] env[63293]: DEBUG nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 743.499037] env[63293]: DEBUG nova.compute.manager [req-6d96cff8-9351-4820-9e12-12d9cde60aa4 req-7fbfd11f-ae07-4375-ba50-99f417f3e336 service nova] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Received event network-changed-37e8a179-ff61-48b6-8d49-44c09262ea80 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 743.499236] env[63293]: DEBUG nova.compute.manager [req-6d96cff8-9351-4820-9e12-12d9cde60aa4 req-7fbfd11f-ae07-4375-ba50-99f417f3e336 service nova] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Refreshing instance network info cache due to event network-changed-37e8a179-ff61-48b6-8d49-44c09262ea80. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 743.499470] env[63293]: DEBUG oslo_concurrency.lockutils [req-6d96cff8-9351-4820-9e12-12d9cde60aa4 req-7fbfd11f-ae07-4375-ba50-99f417f3e336 service nova] Acquiring lock "refresh_cache-263d1329-a81a-4d50-91b6-3927bffe8d4c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.620994] env[63293]: DEBUG nova.network.neutron [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Successfully created port: 0845133e-ee59-4fca-8d15-f336202ab231 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 743.803066] env[63293]: DEBUG nova.network.neutron [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.856985] env[63293]: DEBUG nova.network.neutron [-] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.923075] env[63293]: DEBUG nova.network.neutron [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.992019] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3586d2f4-2832-4c4e-a52c-5b74b1a845ea {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.999036] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14cb0191-3664-4943-8be4-b337fba06559 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.003985] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 744.005630] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 744.041504] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e076901c-ce1c-4950-95ac-3e7391e7d2e6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.052491] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16194dcf-74d5-49d6-aa87-05353d0a368a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.067643] env[63293]: DEBUG nova.compute.provider_tree [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.083238] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquiring lock "93080446-fd45-4281-bf6e-35d1d964f41e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.083655] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Lock "93080446-fd45-4281-bf6e-35d1d964f41e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.134344] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "c26d6a3a-3643-4b52-870a-4e2ed901d908" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.134614] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "c26d6a3a-3643-4b52-870a-4e2ed901d908" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.362626] env[63293]: INFO nova.compute.manager [-] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Took 1.03 seconds to deallocate network for instance. [ 744.429232] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Releasing lock "refresh_cache-263d1329-a81a-4d50-91b6-3927bffe8d4c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.429232] env[63293]: DEBUG nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 744.429232] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.429748] env[63293]: DEBUG oslo_concurrency.lockutils [req-6d96cff8-9351-4820-9e12-12d9cde60aa4 req-7fbfd11f-ae07-4375-ba50-99f417f3e336 service nova] Acquired lock "refresh_cache-263d1329-a81a-4d50-91b6-3927bffe8d4c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.429988] env[63293]: DEBUG nova.network.neutron [req-6d96cff8-9351-4820-9e12-12d9cde60aa4 req-7fbfd11f-ae07-4375-ba50-99f417f3e336 service nova] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Refreshing network info cache for port 37e8a179-ff61-48b6-8d49-44c09262ea80 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 744.431140] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66fb5940-e8a4-4a80-b440-665e28339536 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.442691] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd3cdc3-250b-4c3b-a6b7-3e906f262e0c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.471162] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 263d1329-a81a-4d50-91b6-3927bffe8d4c could not be found. [ 744.471497] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 744.471787] env[63293]: INFO nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 744.472142] env[63293]: DEBUG oslo.service.loopingcall [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.472460] env[63293]: DEBUG nova.compute.manager [-] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.472737] env[63293]: DEBUG nova.network.neutron [-] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 744.493818] env[63293]: DEBUG nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 744.504144] env[63293]: DEBUG nova.network.neutron [-] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.513213] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 744.514123] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Starting heal instance info cache {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 744.514286] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Rebuilding the list of instances to heal {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 744.530456] env[63293]: DEBUG nova.virt.hardware [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 744.531073] env[63293]: DEBUG nova.virt.hardware [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 744.531294] env[63293]: DEBUG nova.virt.hardware [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 744.531512] env[63293]: DEBUG nova.virt.hardware [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 744.531700] env[63293]: DEBUG nova.virt.hardware [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 744.531880] env[63293]: DEBUG nova.virt.hardware [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 744.532145] env[63293]: DEBUG nova.virt.hardware [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 744.532334] env[63293]: DEBUG nova.virt.hardware [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 744.532543] env[63293]: DEBUG nova.virt.hardware [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 744.532744] env[63293]: DEBUG nova.virt.hardware [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 744.532958] env[63293]: DEBUG nova.virt.hardware [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 744.534258] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38500802-d429-4458-92f3-184409c6ade9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.545684] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3f8b12-2879-46c2-9b44-c35378c638c7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.572027] env[63293]: DEBUG nova.scheduler.client.report [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.747306] env[63293]: DEBUG nova.compute.manager [req-046f8404-b32a-4b24-a578-4939a19fe89e req-bfcaf249-dc67-4ee4-9958-233699ab96f9 service nova] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Received event network-changed-0845133e-ee59-4fca-8d15-f336202ab231 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.747495] env[63293]: DEBUG nova.compute.manager [req-046f8404-b32a-4b24-a578-4939a19fe89e req-bfcaf249-dc67-4ee4-9958-233699ab96f9 service nova] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Refreshing instance network info cache due to event network-changed-0845133e-ee59-4fca-8d15-f336202ab231. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 744.747760] env[63293]: DEBUG oslo_concurrency.lockutils [req-046f8404-b32a-4b24-a578-4939a19fe89e req-bfcaf249-dc67-4ee4-9958-233699ab96f9 service nova] Acquiring lock "refresh_cache-fd67940d-9552-42b3-b472-2ce0e7899ea6" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.747840] env[63293]: DEBUG oslo_concurrency.lockutils [req-046f8404-b32a-4b24-a578-4939a19fe89e req-bfcaf249-dc67-4ee4-9958-233699ab96f9 service nova] Acquired lock "refresh_cache-fd67940d-9552-42b3-b472-2ce0e7899ea6" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.748263] env[63293]: DEBUG nova.network.neutron [req-046f8404-b32a-4b24-a578-4939a19fe89e req-bfcaf249-dc67-4ee4-9958-233699ab96f9 service nova] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Refreshing network info cache for port 0845133e-ee59-4fca-8d15-f336202ab231 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 744.870836] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.953571] env[63293]: ERROR nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0845133e-ee59-4fca-8d15-f336202ab231, please check neutron logs for more information. [ 744.953571] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 744.953571] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.953571] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 744.953571] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.953571] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 744.953571] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.953571] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 744.953571] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.953571] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 744.953571] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.953571] env[63293]: ERROR nova.compute.manager raise self.value [ 744.953571] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.953571] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 744.953571] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.953571] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 744.954185] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.954185] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 744.954185] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0845133e-ee59-4fca-8d15-f336202ab231, please check neutron logs for more information. [ 744.954185] env[63293]: ERROR nova.compute.manager [ 744.954185] env[63293]: Traceback (most recent call last): [ 744.954185] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 744.954185] env[63293]: listener.cb(fileno) [ 744.954185] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.954185] env[63293]: result = function(*args, **kwargs) [ 744.954185] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 744.954185] env[63293]: return func(*args, **kwargs) [ 744.954185] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.954185] env[63293]: raise e [ 744.954185] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.954185] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 744.954185] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.954185] env[63293]: created_port_ids = self._update_ports_for_instance( [ 744.954185] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.954185] env[63293]: with excutils.save_and_reraise_exception(): [ 744.954185] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.954185] env[63293]: self.force_reraise() [ 744.954185] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.954185] env[63293]: raise self.value [ 744.954185] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.954185] env[63293]: updated_port = self._update_port( [ 744.954185] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.954185] env[63293]: _ensure_no_port_binding_failure(port) [ 744.954185] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.954185] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 744.955168] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 0845133e-ee59-4fca-8d15-f336202ab231, please check neutron logs for more information. [ 744.955168] env[63293]: Removing descriptor: 18 [ 744.955168] env[63293]: ERROR nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0845133e-ee59-4fca-8d15-f336202ab231, please check neutron logs for more information. [ 744.955168] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Traceback (most recent call last): [ 744.955168] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 744.955168] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] yield resources [ 744.955168] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 744.955168] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] self.driver.spawn(context, instance, image_meta, [ 744.955168] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 744.955168] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.955168] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.955168] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] vm_ref = self.build_virtual_machine(instance, [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] for vif in network_info: [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] return self._sync_wrapper(fn, *args, **kwargs) [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] self.wait() [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] self[:] = self._gt.wait() [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] return self._exit_event.wait() [ 744.955624] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] result = hub.switch() [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] return self.greenlet.switch() [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] result = function(*args, **kwargs) [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] return func(*args, **kwargs) [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] raise e [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] nwinfo = self.network_api.allocate_for_instance( [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.956141] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] created_port_ids = self._update_ports_for_instance( [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] with excutils.save_and_reraise_exception(): [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] self.force_reraise() [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] raise self.value [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] updated_port = self._update_port( [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] _ensure_no_port_binding_failure(port) [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.956792] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] raise exception.PortBindingFailed(port_id=port['id']) [ 744.957232] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] nova.exception.PortBindingFailed: Binding failed for port 0845133e-ee59-4fca-8d15-f336202ab231, please check neutron logs for more information. [ 744.957232] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] [ 744.957232] env[63293]: INFO nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Terminating instance [ 744.964235] env[63293]: DEBUG nova.network.neutron [req-6d96cff8-9351-4820-9e12-12d9cde60aa4 req-7fbfd11f-ae07-4375-ba50-99f417f3e336 service nova] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.964235] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-fd67940d-9552-42b3-b472-2ce0e7899ea6" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.007675] env[63293]: DEBUG nova.network.neutron [-] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.018706] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 745.020019] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 745.020019] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 745.020019] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 745.038653] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "refresh_cache-0aade4da-882a-4342-aefb-07eb306b17ff" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.038737] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquired lock "refresh_cache-0aade4da-882a-4342-aefb-07eb306b17ff" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.038827] env[63293]: DEBUG nova.network.neutron [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Forcefully refreshing network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 745.038971] env[63293]: DEBUG nova.objects.instance [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lazy-loading 'info_cache' on Instance uuid 0aade4da-882a-4342-aefb-07eb306b17ff {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 745.077377] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.624s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.077910] env[63293]: DEBUG nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 745.081399] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.092s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.084646] env[63293]: INFO nova.compute.claims [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.116107] env[63293]: DEBUG nova.network.neutron [req-6d96cff8-9351-4820-9e12-12d9cde60aa4 req-7fbfd11f-ae07-4375-ba50-99f417f3e336 service nova] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.267797] env[63293]: DEBUG nova.network.neutron [req-046f8404-b32a-4b24-a578-4939a19fe89e req-bfcaf249-dc67-4ee4-9958-233699ab96f9 service nova] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.362196] env[63293]: DEBUG nova.network.neutron [req-046f8404-b32a-4b24-a578-4939a19fe89e req-bfcaf249-dc67-4ee4-9958-233699ab96f9 service nova] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.510805] env[63293]: INFO nova.compute.manager [-] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Took 1.04 seconds to deallocate network for instance. [ 745.513290] env[63293]: DEBUG nova.compute.claims [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 745.513466] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.564455] env[63293]: DEBUG nova.compute.manager [req-45812198-4487-44e2-a830-b7b993dd38c5 req-18e1b668-e705-439e-af7f-394a620f424d service nova] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Received event network-vif-deleted-37e8a179-ff61-48b6-8d49-44c09262ea80 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 745.592035] env[63293]: DEBUG nova.compute.utils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.593542] env[63293]: DEBUG nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 745.593872] env[63293]: DEBUG nova.network.neutron [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 745.620111] env[63293]: DEBUG oslo_concurrency.lockutils [req-6d96cff8-9351-4820-9e12-12d9cde60aa4 req-7fbfd11f-ae07-4375-ba50-99f417f3e336 service nova] Releasing lock "refresh_cache-263d1329-a81a-4d50-91b6-3927bffe8d4c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.658474] env[63293]: DEBUG nova.policy [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '52947b5db3ad4d04a5836a6ffa3e990b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0dc7afcafce845cc90b72fcc23702d9e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.865231] env[63293]: DEBUG oslo_concurrency.lockutils [req-046f8404-b32a-4b24-a578-4939a19fe89e req-bfcaf249-dc67-4ee4-9958-233699ab96f9 service nova] Releasing lock "refresh_cache-fd67940d-9552-42b3-b472-2ce0e7899ea6" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.865645] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-fd67940d-9552-42b3-b472-2ce0e7899ea6" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.865881] env[63293]: DEBUG nova.network.neutron [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.987823] env[63293]: DEBUG nova.network.neutron [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Successfully created port: dbed6212-8b1d-4e39-bcd7-b0cf04b60660 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.073099] env[63293]: DEBUG nova.network.neutron [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.094262] env[63293]: DEBUG nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 746.276711] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "9ddf4f23-1279-4cbd-8212-10f344060445" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.276931] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.389780] env[63293]: DEBUG nova.network.neutron [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.494355] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250ca4ad-d79c-4f9c-8fdb-679cf031823f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.502431] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2b805a-179a-490c-99e2-0ed7a0afd190 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.534007] env[63293]: DEBUG nova.network.neutron [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.538642] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c870925-7b8f-4cdc-9f94-26f33bccca3f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.545041] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3a50d3-9193-4cde-86e1-a259c1e226cb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.560033] env[63293]: DEBUG nova.compute.provider_tree [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.626700] env[63293]: DEBUG nova.network.neutron [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.777746] env[63293]: DEBUG nova.compute.manager [req-d1994698-5660-4bac-81ad-c387aab98d0a req-1d7de1ad-32e4-44d2-a9f7-c5736061bbff service nova] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Received event network-vif-deleted-0845133e-ee59-4fca-8d15-f336202ab231 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 747.005452] env[63293]: ERROR nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dbed6212-8b1d-4e39-bcd7-b0cf04b60660, please check neutron logs for more information. [ 747.005452] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 747.005452] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.005452] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 747.005452] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.005452] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 747.005452] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.005452] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 747.005452] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.005452] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 747.005452] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.005452] env[63293]: ERROR nova.compute.manager raise self.value [ 747.005452] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.005452] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 747.005452] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.005452] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 747.006321] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.006321] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 747.006321] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dbed6212-8b1d-4e39-bcd7-b0cf04b60660, please check neutron logs for more information. [ 747.006321] env[63293]: ERROR nova.compute.manager [ 747.006321] env[63293]: Traceback (most recent call last): [ 747.006321] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 747.006321] env[63293]: listener.cb(fileno) [ 747.006321] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.006321] env[63293]: result = function(*args, **kwargs) [ 747.006321] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.006321] env[63293]: return func(*args, **kwargs) [ 747.006321] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.006321] env[63293]: raise e [ 747.006321] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.006321] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 747.006321] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.006321] env[63293]: created_port_ids = self._update_ports_for_instance( [ 747.006321] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.006321] env[63293]: with excutils.save_and_reraise_exception(): [ 747.006321] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.006321] env[63293]: self.force_reraise() [ 747.006321] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.006321] env[63293]: raise self.value [ 747.006321] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.006321] env[63293]: updated_port = self._update_port( [ 747.006321] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.006321] env[63293]: _ensure_no_port_binding_failure(port) [ 747.006321] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.006321] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 747.007671] env[63293]: nova.exception.PortBindingFailed: Binding failed for port dbed6212-8b1d-4e39-bcd7-b0cf04b60660, please check neutron logs for more information. [ 747.007671] env[63293]: Removing descriptor: 15 [ 747.040622] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-fd67940d-9552-42b3-b472-2ce0e7899ea6" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.041032] env[63293]: DEBUG nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.041187] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.041471] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8116c20-c8c1-4347-8d3f-43ef47a7bf89 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.051560] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c7e53d-bf37-4cd9-85c8-f65159b30a88 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.062224] env[63293]: DEBUG nova.scheduler.client.report [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.072396] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fd67940d-9552-42b3-b472-2ce0e7899ea6 could not be found. [ 747.072597] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 747.072773] env[63293]: INFO nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 747.073013] env[63293]: DEBUG oslo.service.loopingcall [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.073238] env[63293]: DEBUG nova.compute.manager [-] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.073360] env[63293]: DEBUG nova.network.neutron [-] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 747.089222] env[63293]: DEBUG nova.network.neutron [-] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.107348] env[63293]: DEBUG nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 747.128963] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Releasing lock "refresh_cache-0aade4da-882a-4342-aefb-07eb306b17ff" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.129293] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Updated the network info_cache for instance {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 747.129760] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 747.129964] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 747.132230] env[63293]: DEBUG nova.virt.hardware [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 747.132395] env[63293]: DEBUG nova.virt.hardware [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 747.132467] env[63293]: DEBUG nova.virt.hardware [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.132588] env[63293]: DEBUG nova.virt.hardware [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 747.132736] env[63293]: DEBUG nova.virt.hardware [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.132881] env[63293]: DEBUG nova.virt.hardware [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 747.133096] env[63293]: DEBUG nova.virt.hardware [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 747.133246] env[63293]: DEBUG nova.virt.hardware [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 747.133412] env[63293]: DEBUG nova.virt.hardware [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 747.133604] env[63293]: DEBUG nova.virt.hardware [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 747.133723] env[63293]: DEBUG nova.virt.hardware [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 747.133973] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 747.134730] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569067cc-eb11-4452-9fe9-140bd4e2be49 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.137454] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 747.137625] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 747.138164] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 747.138373] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63293) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 747.138588] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 747.144623] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93bf06cc-1cad-428f-9186-d62a8d52bce0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.174378] env[63293]: ERROR nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dbed6212-8b1d-4e39-bcd7-b0cf04b60660, please check neutron logs for more information. [ 747.174378] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Traceback (most recent call last): [ 747.174378] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 747.174378] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] yield resources [ 747.174378] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 747.174378] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] self.driver.spawn(context, instance, image_meta, [ 747.174378] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 747.174378] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.174378] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.174378] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] vm_ref = self.build_virtual_machine(instance, [ 747.174378] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] for vif in network_info: [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] return self._sync_wrapper(fn, *args, **kwargs) [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] self.wait() [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] self[:] = self._gt.wait() [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] return self._exit_event.wait() [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 747.175063] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] current.throw(*self._exc) [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] result = function(*args, **kwargs) [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] return func(*args, **kwargs) [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] raise e [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] nwinfo = self.network_api.allocate_for_instance( [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] created_port_ids = self._update_ports_for_instance( [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] with excutils.save_and_reraise_exception(): [ 747.175675] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.176307] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] self.force_reraise() [ 747.176307] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.176307] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] raise self.value [ 747.176307] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.176307] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] updated_port = self._update_port( [ 747.176307] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.176307] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] _ensure_no_port_binding_failure(port) [ 747.176307] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.176307] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] raise exception.PortBindingFailed(port_id=port['id']) [ 747.176307] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] nova.exception.PortBindingFailed: Binding failed for port dbed6212-8b1d-4e39-bcd7-b0cf04b60660, please check neutron logs for more information. [ 747.176307] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] [ 747.176307] env[63293]: INFO nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Terminating instance [ 747.179092] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Acquiring lock "refresh_cache-1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.179209] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Acquired lock "refresh_cache-1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.179455] env[63293]: DEBUG nova.network.neutron [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.567218] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.567796] env[63293]: DEBUG nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 747.570552] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.650s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.573208] env[63293]: INFO nova.compute.claims [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.591213] env[63293]: DEBUG nova.network.neutron [-] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.596397] env[63293]: DEBUG nova.compute.manager [req-e41bbc00-a8fe-4f2a-b500-ed434d4d9e24 req-ef122004-977e-4a66-9e07-52af5310c215 service nova] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Received event network-changed-dbed6212-8b1d-4e39-bcd7-b0cf04b60660 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 747.596638] env[63293]: DEBUG nova.compute.manager [req-e41bbc00-a8fe-4f2a-b500-ed434d4d9e24 req-ef122004-977e-4a66-9e07-52af5310c215 service nova] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Refreshing instance network info cache due to event network-changed-dbed6212-8b1d-4e39-bcd7-b0cf04b60660. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 747.597366] env[63293]: DEBUG oslo_concurrency.lockutils [req-e41bbc00-a8fe-4f2a-b500-ed434d4d9e24 req-ef122004-977e-4a66-9e07-52af5310c215 service nova] Acquiring lock "refresh_cache-1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.642357] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.697464] env[63293]: DEBUG nova.network.neutron [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.787370] env[63293]: DEBUG nova.network.neutron [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.077241] env[63293]: DEBUG nova.compute.utils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.082380] env[63293]: DEBUG nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.082874] env[63293]: DEBUG nova.network.neutron [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 748.098497] env[63293]: INFO nova.compute.manager [-] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Took 1.02 seconds to deallocate network for instance. [ 748.100789] env[63293]: DEBUG nova.compute.claims [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 748.100965] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.132655] env[63293]: DEBUG nova.policy [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b02cabad3b564ff09b26abe1dfde8ad4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bcb70627939d4addb756ccbc409fd07f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 748.289978] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Releasing lock "refresh_cache-1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.290438] env[63293]: DEBUG nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 748.290631] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 748.290960] env[63293]: DEBUG oslo_concurrency.lockutils [req-e41bbc00-a8fe-4f2a-b500-ed434d4d9e24 req-ef122004-977e-4a66-9e07-52af5310c215 service nova] Acquired lock "refresh_cache-1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.291150] env[63293]: DEBUG nova.network.neutron [req-e41bbc00-a8fe-4f2a-b500-ed434d4d9e24 req-ef122004-977e-4a66-9e07-52af5310c215 service nova] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Refreshing network info cache for port dbed6212-8b1d-4e39-bcd7-b0cf04b60660 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 748.292646] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac09f68e-c34a-4e16-9135-d9987780bc0f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.302596] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714170a5-6f9c-4cd8-8fed-e1590680094c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.326085] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9 could not be found. [ 748.326317] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.326504] env[63293]: INFO nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 748.326740] env[63293]: DEBUG oslo.service.loopingcall [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.326952] env[63293]: DEBUG nova.compute.manager [-] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.327061] env[63293]: DEBUG nova.network.neutron [-] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.346680] env[63293]: DEBUG nova.network.neutron [-] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.425718] env[63293]: DEBUG nova.network.neutron [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Successfully created port: 3b1821c8-1889-473b-8f69-e57fa4768f18 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.586365] env[63293]: DEBUG nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 748.816603] env[63293]: DEBUG nova.network.neutron [req-e41bbc00-a8fe-4f2a-b500-ed434d4d9e24 req-ef122004-977e-4a66-9e07-52af5310c215 service nova] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.849542] env[63293]: DEBUG nova.network.neutron [-] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.913809] env[63293]: DEBUG nova.network.neutron [req-e41bbc00-a8fe-4f2a-b500-ed434d4d9e24 req-ef122004-977e-4a66-9e07-52af5310c215 service nova] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.940203] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e5fe5f-ff16-4f27-8c5d-4811967cef78 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.948101] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5c7c88-cb83-4f89-85f3-98b50b6a5a9d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.983604] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf10e47-e80b-4ce2-b7f5-2a0fc479817c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.992257] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5d321d-36b6-4631-a618-0a73f4400703 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.007111] env[63293]: DEBUG nova.compute.provider_tree [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.352029] env[63293]: INFO nova.compute.manager [-] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Took 1.02 seconds to deallocate network for instance. [ 749.354507] env[63293]: DEBUG nova.compute.claims [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 749.354634] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.416072] env[63293]: DEBUG oslo_concurrency.lockutils [req-e41bbc00-a8fe-4f2a-b500-ed434d4d9e24 req-ef122004-977e-4a66-9e07-52af5310c215 service nova] Releasing lock "refresh_cache-1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.416359] env[63293]: DEBUG nova.compute.manager [req-e41bbc00-a8fe-4f2a-b500-ed434d4d9e24 req-ef122004-977e-4a66-9e07-52af5310c215 service nova] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Received event network-vif-deleted-dbed6212-8b1d-4e39-bcd7-b0cf04b60660 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.428233] env[63293]: ERROR nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3b1821c8-1889-473b-8f69-e57fa4768f18, please check neutron logs for more information. [ 749.428233] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 749.428233] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.428233] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 749.428233] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.428233] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 749.428233] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.428233] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 749.428233] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.428233] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 749.428233] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.428233] env[63293]: ERROR nova.compute.manager raise self.value [ 749.428233] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.428233] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 749.428233] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.428233] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 749.429217] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.429217] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 749.429217] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3b1821c8-1889-473b-8f69-e57fa4768f18, please check neutron logs for more information. [ 749.429217] env[63293]: ERROR nova.compute.manager [ 749.429217] env[63293]: Traceback (most recent call last): [ 749.429217] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 749.429217] env[63293]: listener.cb(fileno) [ 749.429217] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.429217] env[63293]: result = function(*args, **kwargs) [ 749.429217] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.429217] env[63293]: return func(*args, **kwargs) [ 749.429217] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.429217] env[63293]: raise e [ 749.429217] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.429217] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 749.429217] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.429217] env[63293]: created_port_ids = self._update_ports_for_instance( [ 749.429217] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.429217] env[63293]: with excutils.save_and_reraise_exception(): [ 749.429217] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.429217] env[63293]: self.force_reraise() [ 749.429217] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.429217] env[63293]: raise self.value [ 749.429217] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.429217] env[63293]: updated_port = self._update_port( [ 749.429217] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.429217] env[63293]: _ensure_no_port_binding_failure(port) [ 749.429217] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.429217] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 749.430660] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 3b1821c8-1889-473b-8f69-e57fa4768f18, please check neutron logs for more information. [ 749.430660] env[63293]: Removing descriptor: 15 [ 749.512440] env[63293]: DEBUG nova.scheduler.client.report [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.599307] env[63293]: DEBUG nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 749.622340] env[63293]: DEBUG nova.compute.manager [req-d9fb8dc0-54dc-4fe0-91ea-7765d8165d0c req-4bc1f793-45cb-41cc-9333-4403bcdf8415 service nova] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Received event network-changed-3b1821c8-1889-473b-8f69-e57fa4768f18 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.622533] env[63293]: DEBUG nova.compute.manager [req-d9fb8dc0-54dc-4fe0-91ea-7765d8165d0c req-4bc1f793-45cb-41cc-9333-4403bcdf8415 service nova] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Refreshing instance network info cache due to event network-changed-3b1821c8-1889-473b-8f69-e57fa4768f18. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 749.622742] env[63293]: DEBUG oslo_concurrency.lockutils [req-d9fb8dc0-54dc-4fe0-91ea-7765d8165d0c req-4bc1f793-45cb-41cc-9333-4403bcdf8415 service nova] Acquiring lock "refresh_cache-985a5c30-551f-4de7-bc85-d53f002e0f72" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.622880] env[63293]: DEBUG oslo_concurrency.lockutils [req-d9fb8dc0-54dc-4fe0-91ea-7765d8165d0c req-4bc1f793-45cb-41cc-9333-4403bcdf8415 service nova] Acquired lock "refresh_cache-985a5c30-551f-4de7-bc85-d53f002e0f72" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.623041] env[63293]: DEBUG nova.network.neutron [req-d9fb8dc0-54dc-4fe0-91ea-7765d8165d0c req-4bc1f793-45cb-41cc-9333-4403bcdf8415 service nova] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Refreshing network info cache for port 3b1821c8-1889-473b-8f69-e57fa4768f18 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 749.625831] env[63293]: DEBUG nova.virt.hardware [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.626057] env[63293]: DEBUG nova.virt.hardware [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.626217] env[63293]: DEBUG nova.virt.hardware [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.626406] env[63293]: DEBUG nova.virt.hardware [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.626539] env[63293]: DEBUG nova.virt.hardware [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.626684] env[63293]: DEBUG nova.virt.hardware [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.626884] env[63293]: DEBUG nova.virt.hardware [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.627053] env[63293]: DEBUG nova.virt.hardware [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.627222] env[63293]: DEBUG nova.virt.hardware [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.627381] env[63293]: DEBUG nova.virt.hardware [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.627548] env[63293]: DEBUG nova.virt.hardware [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.629129] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5efd110e-17b6-4539-b4e7-ac1175079eed {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.638802] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e785a47-3e16-4fc3-be47-471e15facb28 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.653600] env[63293]: ERROR nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3b1821c8-1889-473b-8f69-e57fa4768f18, please check neutron logs for more information. [ 749.653600] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Traceback (most recent call last): [ 749.653600] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 749.653600] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] yield resources [ 749.653600] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.653600] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] self.driver.spawn(context, instance, image_meta, [ 749.653600] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 749.653600] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.653600] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.653600] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] vm_ref = self.build_virtual_machine(instance, [ 749.653600] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] for vif in network_info: [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] return self._sync_wrapper(fn, *args, **kwargs) [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] self.wait() [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] self[:] = self._gt.wait() [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] return self._exit_event.wait() [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 749.653996] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] current.throw(*self._exc) [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] result = function(*args, **kwargs) [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] return func(*args, **kwargs) [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] raise e [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] nwinfo = self.network_api.allocate_for_instance( [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] created_port_ids = self._update_ports_for_instance( [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] with excutils.save_and_reraise_exception(): [ 749.654695] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.655356] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] self.force_reraise() [ 749.655356] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.655356] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] raise self.value [ 749.655356] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.655356] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] updated_port = self._update_port( [ 749.655356] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.655356] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] _ensure_no_port_binding_failure(port) [ 749.655356] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.655356] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] raise exception.PortBindingFailed(port_id=port['id']) [ 749.655356] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] nova.exception.PortBindingFailed: Binding failed for port 3b1821c8-1889-473b-8f69-e57fa4768f18, please check neutron logs for more information. [ 749.655356] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] [ 749.655356] env[63293]: INFO nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Terminating instance [ 749.656128] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Acquiring lock "refresh_cache-985a5c30-551f-4de7-bc85-d53f002e0f72" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.017602] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.018328] env[63293]: DEBUG nova.compute.manager [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 750.022113] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.575s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.024138] env[63293]: INFO nova.compute.claims [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.149443] env[63293]: DEBUG nova.network.neutron [req-d9fb8dc0-54dc-4fe0-91ea-7765d8165d0c req-4bc1f793-45cb-41cc-9333-4403bcdf8415 service nova] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.226868] env[63293]: DEBUG nova.network.neutron [req-d9fb8dc0-54dc-4fe0-91ea-7765d8165d0c req-4bc1f793-45cb-41cc-9333-4403bcdf8415 service nova] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.527808] env[63293]: DEBUG nova.compute.utils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 750.531777] env[63293]: DEBUG nova.compute.manager [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Not allocating networking since 'none' was specified. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 750.726278] env[63293]: DEBUG oslo_concurrency.lockutils [req-d9fb8dc0-54dc-4fe0-91ea-7765d8165d0c req-4bc1f793-45cb-41cc-9333-4403bcdf8415 service nova] Releasing lock "refresh_cache-985a5c30-551f-4de7-bc85-d53f002e0f72" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.727033] env[63293]: DEBUG nova.compute.manager [req-d9fb8dc0-54dc-4fe0-91ea-7765d8165d0c req-4bc1f793-45cb-41cc-9333-4403bcdf8415 service nova] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Received event network-vif-deleted-3b1821c8-1889-473b-8f69-e57fa4768f18 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.727252] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Acquired lock "refresh_cache-985a5c30-551f-4de7-bc85-d53f002e0f72" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.727488] env[63293]: DEBUG nova.network.neutron [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.032098] env[63293]: DEBUG nova.compute.manager [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 751.245898] env[63293]: DEBUG nova.network.neutron [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.315751] env[63293]: DEBUG nova.network.neutron [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.342973] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a4a9d3-973a-485b-aeac-a5896d410346 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.352857] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b8d413-28d5-4dc3-9d78-8460eb9bd717 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.383496] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b846c227-cdb4-4a84-aca5-a0a83182249f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.392346] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04dcc7aa-d447-4fc1-86ce-72e6d938f8d9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.405061] env[63293]: DEBUG nova.compute.provider_tree [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.818863] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Releasing lock "refresh_cache-985a5c30-551f-4de7-bc85-d53f002e0f72" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.819158] env[63293]: DEBUG nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.819158] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.819530] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eaed5d1c-ec8f-45a4-8b42-9bdab362d456 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.828808] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefd5701-e303-4e23-8b09-c29a42753c49 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.850091] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 985a5c30-551f-4de7-bc85-d53f002e0f72 could not be found. [ 751.850310] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 751.850476] env[63293]: INFO nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Took 0.03 seconds to destroy the instance on the hypervisor. [ 751.850711] env[63293]: DEBUG oslo.service.loopingcall [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.850916] env[63293]: DEBUG nova.compute.manager [-] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.851014] env[63293]: DEBUG nova.network.neutron [-] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 751.865276] env[63293]: DEBUG nova.network.neutron [-] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.907673] env[63293]: DEBUG nova.scheduler.client.report [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.044825] env[63293]: DEBUG nova.compute.manager [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 752.068927] env[63293]: DEBUG nova.virt.hardware [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.069175] env[63293]: DEBUG nova.virt.hardware [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.069331] env[63293]: DEBUG nova.virt.hardware [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.069525] env[63293]: DEBUG nova.virt.hardware [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.069674] env[63293]: DEBUG nova.virt.hardware [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.069818] env[63293]: DEBUG nova.virt.hardware [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.070081] env[63293]: DEBUG nova.virt.hardware [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.070935] env[63293]: DEBUG nova.virt.hardware [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.070935] env[63293]: DEBUG nova.virt.hardware [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.070935] env[63293]: DEBUG nova.virt.hardware [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.070935] env[63293]: DEBUG nova.virt.hardware [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.071617] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa449b7c-1c7d-4e36-9450-e587443707de {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.079812] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a0b940-0a27-400c-80c5-ec068625d61b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.092751] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 752.098189] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Creating folder: Project (297b494b75424de9a7e1657a1a9a10b2). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 752.098496] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-249ac71f-5686-491c-b251-71542144f35d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.108068] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Created folder: Project (297b494b75424de9a7e1657a1a9a10b2) in parent group-v283678. [ 752.108068] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Creating folder: Instances. Parent ref: group-v283698. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 752.108221] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4805e123-3cc2-46e8-99f0-42f4957d5ca7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.116920] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Created folder: Instances in parent group-v283698. [ 752.117131] env[63293]: DEBUG oslo.service.loopingcall [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.117262] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5334832-25dc-4b03-9689-388a1382200a] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 752.117570] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c06f787b-eb51-4e39-8262-8d38517997c6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.136232] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 752.136232] env[63293]: value = "task-1327487" [ 752.136232] env[63293]: _type = "Task" [ 752.136232] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.148551] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327487, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.369198] env[63293]: DEBUG nova.network.neutron [-] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.413690] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.414312] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 752.417374] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.961s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.418954] env[63293]: INFO nova.compute.claims [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.646130] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327487, 'name': CreateVM_Task, 'duration_secs': 0.370129} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.646700] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5334832-25dc-4b03-9689-388a1382200a] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 752.646784] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.646936] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.647256] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 752.647491] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16773b02-0f4a-4201-89b4-991656e8a5d4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.651659] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 752.651659] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52cd75e1-3b76-3e78-9b60-622f591ce8fd" [ 752.651659] env[63293]: _type = "Task" [ 752.651659] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.659610] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52cd75e1-3b76-3e78-9b60-622f591ce8fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.873597] env[63293]: INFO nova.compute.manager [-] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Took 1.02 seconds to deallocate network for instance. [ 752.879375] env[63293]: DEBUG nova.compute.claims [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 752.879619] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.925766] env[63293]: DEBUG nova.compute.utils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 752.928758] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 752.928758] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 752.978533] env[63293]: DEBUG nova.policy [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c34e72ea18dd409ba5d40b32711f456f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49f167005da748fcaf2497a5040a838c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 753.161697] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52cd75e1-3b76-3e78-9b60-622f591ce8fd, 'name': SearchDatastore_Task, 'duration_secs': 0.033466} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.161995] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.162237] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 753.162468] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.162612] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.162782] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 753.163038] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1fe5cae5-7505-470b-9ab9-83b6923b7338 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.172054] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 753.172237] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 753.172918] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a449de5d-6cfb-4fa4-adb6-aa81e2b4221b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.177830] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 753.177830] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c67c48-5ab0-c4a9-529c-1f2ab50a57a1" [ 753.177830] env[63293]: _type = "Task" [ 753.177830] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.184872] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c67c48-5ab0-c4a9-529c-1f2ab50a57a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.246460] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Successfully created port: 536be720-3b98-49a5-a4c9-cfa83b018886 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 753.430455] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 753.688280] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c67c48-5ab0-c4a9-529c-1f2ab50a57a1, 'name': SearchDatastore_Task, 'duration_secs': 0.059343} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.689184] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d223894-730c-46fb-88bb-b67ef98cd4a3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.694299] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 753.694299] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a0b7d9-093b-2687-9e4c-576eb9c64834" [ 753.694299] env[63293]: _type = "Task" [ 753.694299] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.701828] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a0b7d9-093b-2687-9e4c-576eb9c64834, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.823138] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac79611-5923-4ac1-a773-43daee561f6e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.830952] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f759b6-72b3-4e92-8580-62728de33667 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.861380] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c766589-a289-4aea-bf01-ddfff92d601d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.868726] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9035e70-6505-4291-9632-5d042462244e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.882590] env[63293]: DEBUG nova.compute.provider_tree [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.048131] env[63293]: DEBUG nova.compute.manager [req-7d4ce943-1a40-4c8d-a021-7ffa83097a64 req-6c7c8a08-5ef3-4398-9602-3a2df28ad21c service nova] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Received event network-changed-536be720-3b98-49a5-a4c9-cfa83b018886 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.048339] env[63293]: DEBUG nova.compute.manager [req-7d4ce943-1a40-4c8d-a021-7ffa83097a64 req-6c7c8a08-5ef3-4398-9602-3a2df28ad21c service nova] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Refreshing instance network info cache due to event network-changed-536be720-3b98-49a5-a4c9-cfa83b018886. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 754.048630] env[63293]: DEBUG oslo_concurrency.lockutils [req-7d4ce943-1a40-4c8d-a021-7ffa83097a64 req-6c7c8a08-5ef3-4398-9602-3a2df28ad21c service nova] Acquiring lock "refresh_cache-67839a84-c1a0-4728-88f4-47acf99f1fc4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.048825] env[63293]: DEBUG oslo_concurrency.lockutils [req-7d4ce943-1a40-4c8d-a021-7ffa83097a64 req-6c7c8a08-5ef3-4398-9602-3a2df28ad21c service nova] Acquired lock "refresh_cache-67839a84-c1a0-4728-88f4-47acf99f1fc4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.049017] env[63293]: DEBUG nova.network.neutron [req-7d4ce943-1a40-4c8d-a021-7ffa83097a64 req-6c7c8a08-5ef3-4398-9602-3a2df28ad21c service nova] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Refreshing network info cache for port 536be720-3b98-49a5-a4c9-cfa83b018886 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 754.205824] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a0b7d9-093b-2687-9e4c-576eb9c64834, 'name': SearchDatastore_Task, 'duration_secs': 0.01055} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.205824] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.205824] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] b5334832-25dc-4b03-9689-388a1382200a/b5334832-25dc-4b03-9689-388a1382200a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 754.205824] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dacddca0-000a-4aec-816e-23411a318d1a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.212186] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 754.212186] env[63293]: value = "task-1327488" [ 754.212186] env[63293]: _type = "Task" [ 754.212186] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.220315] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327488, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.240964] env[63293]: ERROR nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 536be720-3b98-49a5-a4c9-cfa83b018886, please check neutron logs for more information. [ 754.240964] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 754.240964] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.240964] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 754.240964] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.240964] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 754.240964] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.240964] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 754.240964] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.240964] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 754.240964] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.240964] env[63293]: ERROR nova.compute.manager raise self.value [ 754.240964] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.240964] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 754.240964] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.240964] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 754.241733] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.241733] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 754.241733] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 536be720-3b98-49a5-a4c9-cfa83b018886, please check neutron logs for more information. [ 754.241733] env[63293]: ERROR nova.compute.manager [ 754.241733] env[63293]: Traceback (most recent call last): [ 754.241733] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 754.241733] env[63293]: listener.cb(fileno) [ 754.241733] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.241733] env[63293]: result = function(*args, **kwargs) [ 754.241733] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 754.241733] env[63293]: return func(*args, **kwargs) [ 754.241733] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.241733] env[63293]: raise e [ 754.241733] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.241733] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 754.241733] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.241733] env[63293]: created_port_ids = self._update_ports_for_instance( [ 754.241733] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.241733] env[63293]: with excutils.save_and_reraise_exception(): [ 754.241733] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.241733] env[63293]: self.force_reraise() [ 754.241733] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.241733] env[63293]: raise self.value [ 754.241733] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.241733] env[63293]: updated_port = self._update_port( [ 754.241733] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.241733] env[63293]: _ensure_no_port_binding_failure(port) [ 754.241733] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.241733] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 754.243144] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 536be720-3b98-49a5-a4c9-cfa83b018886, please check neutron logs for more information. [ 754.243144] env[63293]: Removing descriptor: 15 [ 754.386460] env[63293]: DEBUG nova.scheduler.client.report [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.442075] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 754.469882] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 754.470150] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 754.470307] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 754.470487] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 754.470688] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 754.470871] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 754.471096] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 754.471260] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 754.471422] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 754.471583] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 754.471748] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 754.473030] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef68e676-627e-4049-be1f-ad90fdd976ed {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.481307] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47d47e9-3554-420d-a63b-30bf9b5557f2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.496031] env[63293]: ERROR nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 536be720-3b98-49a5-a4c9-cfa83b018886, please check neutron logs for more information. [ 754.496031] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Traceback (most recent call last): [ 754.496031] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 754.496031] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] yield resources [ 754.496031] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 754.496031] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] self.driver.spawn(context, instance, image_meta, [ 754.496031] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 754.496031] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 754.496031] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 754.496031] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] vm_ref = self.build_virtual_machine(instance, [ 754.496031] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] vif_infos = vmwarevif.get_vif_info(self._session, [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] for vif in network_info: [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] return self._sync_wrapper(fn, *args, **kwargs) [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] self.wait() [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] self[:] = self._gt.wait() [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] return self._exit_event.wait() [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 754.496447] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] current.throw(*self._exc) [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] result = function(*args, **kwargs) [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] return func(*args, **kwargs) [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] raise e [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] nwinfo = self.network_api.allocate_for_instance( [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] created_port_ids = self._update_ports_for_instance( [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] with excutils.save_and_reraise_exception(): [ 754.496930] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.497420] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] self.force_reraise() [ 754.497420] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.497420] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] raise self.value [ 754.497420] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.497420] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] updated_port = self._update_port( [ 754.497420] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.497420] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] _ensure_no_port_binding_failure(port) [ 754.497420] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.497420] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] raise exception.PortBindingFailed(port_id=port['id']) [ 754.497420] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] nova.exception.PortBindingFailed: Binding failed for port 536be720-3b98-49a5-a4c9-cfa83b018886, please check neutron logs for more information. [ 754.497420] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] [ 754.497420] env[63293]: INFO nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Terminating instance [ 754.498566] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "refresh_cache-67839a84-c1a0-4728-88f4-47acf99f1fc4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.569881] env[63293]: DEBUG nova.network.neutron [req-7d4ce943-1a40-4c8d-a021-7ffa83097a64 req-6c7c8a08-5ef3-4398-9602-3a2df28ad21c service nova] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.668868] env[63293]: DEBUG nova.network.neutron [req-7d4ce943-1a40-4c8d-a021-7ffa83097a64 req-6c7c8a08-5ef3-4398-9602-3a2df28ad21c service nova] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.722489] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327488, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.892484] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.892484] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 754.895068] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.242s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.896444] env[63293]: INFO nova.compute.claims [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.172996] env[63293]: DEBUG oslo_concurrency.lockutils [req-7d4ce943-1a40-4c8d-a021-7ffa83097a64 req-6c7c8a08-5ef3-4398-9602-3a2df28ad21c service nova] Releasing lock "refresh_cache-67839a84-c1a0-4728-88f4-47acf99f1fc4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.173120] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired lock "refresh_cache-67839a84-c1a0-4728-88f4-47acf99f1fc4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.173256] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 755.223030] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327488, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545876} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.223030] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] b5334832-25dc-4b03-9689-388a1382200a/b5334832-25dc-4b03-9689-388a1382200a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 755.223228] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 755.223437] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b3bbd417-c96b-431b-a4e0-0b164c169822 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.229379] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 755.229379] env[63293]: value = "task-1327489" [ 755.229379] env[63293]: _type = "Task" [ 755.229379] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.236773] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327489, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.401394] env[63293]: DEBUG nova.compute.utils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.404320] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.404484] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 755.452712] env[63293]: DEBUG nova.policy [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c34e72ea18dd409ba5d40b32711f456f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49f167005da748fcaf2497a5040a838c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 755.700322] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.740739] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327489, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073067} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.741011] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.741767] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50da6c55-a403-4610-82d4-f788d5267c61 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.760966] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] b5334832-25dc-4b03-9689-388a1382200a/b5334832-25dc-4b03-9689-388a1382200a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.763756] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Successfully created port: 861f3155-fa0f-452c-8005-af67c9fae028 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 755.765292] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-727123ea-b558-491a-9361-5dd1d052637e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.785856] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 755.785856] env[63293]: value = "task-1327490" [ 755.785856] env[63293]: _type = "Task" [ 755.785856] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.793678] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327490, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.820204] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.905452] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 756.108932] env[63293]: DEBUG nova.compute.manager [req-3a6e0f8b-268b-43ea-9fd7-5d9db0bd18a2 req-aee3f081-79c0-47c0-9f2c-cf768981b69f service nova] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Received event network-vif-deleted-536be720-3b98-49a5-a4c9-cfa83b018886 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 756.288265] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79621dd-e2b6-44f0-a3f3-5e55a5046a59 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.301551] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84ff8e3-9053-4364-b306-d6ee5ec0648d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.304586] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327490, 'name': ReconfigVM_Task, 'duration_secs': 0.290859} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.304845] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Reconfigured VM instance instance-00000030 to attach disk [datastore1] b5334832-25dc-4b03-9689-388a1382200a/b5334832-25dc-4b03-9689-388a1382200a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 756.305734] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa092a3f-4183-4b79-893d-c034121d2251 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.331753] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Releasing lock "refresh_cache-67839a84-c1a0-4728-88f4-47acf99f1fc4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.332332] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 756.332482] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 756.334159] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa789572-0875-48ba-b879-1c4e3d40498e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.336622] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f6c712-e5fe-4e40-9b05-995bf010c681 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.339281] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 756.339281] env[63293]: value = "task-1327491" [ 756.339281] env[63293]: _type = "Task" [ 756.339281] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.350277] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edb0c13-5b82-4efc-8aba-2ddad0279d84 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.361375] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133000cb-476c-4607-ace6-33eda118baaf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.368330] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327491, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.378255] env[63293]: DEBUG nova.compute.provider_tree [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.384362] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 67839a84-c1a0-4728-88f4-47acf99f1fc4 could not be found. [ 756.384579] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 756.384771] env[63293]: INFO nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 756.385032] env[63293]: DEBUG oslo.service.loopingcall [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 756.385466] env[63293]: DEBUG nova.compute.manager [-] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.385565] env[63293]: DEBUG nova.network.neutron [-] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 756.400898] env[63293]: DEBUG nova.network.neutron [-] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.810175] env[63293]: ERROR nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 861f3155-fa0f-452c-8005-af67c9fae028, please check neutron logs for more information. [ 756.810175] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 756.810175] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.810175] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 756.810175] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.810175] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 756.810175] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.810175] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 756.810175] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.810175] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 756.810175] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.810175] env[63293]: ERROR nova.compute.manager raise self.value [ 756.810175] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.810175] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 756.810175] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.810175] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 756.810648] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.810648] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 756.810648] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 861f3155-fa0f-452c-8005-af67c9fae028, please check neutron logs for more information. [ 756.810648] env[63293]: ERROR nova.compute.manager [ 756.810648] env[63293]: Traceback (most recent call last): [ 756.810648] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 756.810648] env[63293]: listener.cb(fileno) [ 756.810648] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.810648] env[63293]: result = function(*args, **kwargs) [ 756.810648] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.810648] env[63293]: return func(*args, **kwargs) [ 756.810648] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.810648] env[63293]: raise e [ 756.810648] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.810648] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 756.810648] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.810648] env[63293]: created_port_ids = self._update_ports_for_instance( [ 756.810648] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.810648] env[63293]: with excutils.save_and_reraise_exception(): [ 756.810648] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.810648] env[63293]: self.force_reraise() [ 756.810648] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.810648] env[63293]: raise self.value [ 756.810648] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.810648] env[63293]: updated_port = self._update_port( [ 756.810648] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.810648] env[63293]: _ensure_no_port_binding_failure(port) [ 756.810648] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.810648] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 756.811578] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 861f3155-fa0f-452c-8005-af67c9fae028, please check neutron logs for more information. [ 756.811578] env[63293]: Removing descriptor: 15 [ 756.855257] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327491, 'name': Rename_Task, 'duration_secs': 0.130769} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.855257] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 756.855257] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee09c3b7-ff88-41f1-8ca5-e6c4684336b9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.859807] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 756.859807] env[63293]: value = "task-1327492" [ 756.859807] env[63293]: _type = "Task" [ 756.859807] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.868428] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327492, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.886678] env[63293]: DEBUG nova.scheduler.client.report [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.903270] env[63293]: DEBUG nova.network.neutron [-] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.918776] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 756.956201] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 756.956441] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 756.956594] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.956770] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 756.956914] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.957076] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 756.957287] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 756.957444] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 756.957654] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 756.957770] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 756.957936] env[63293]: DEBUG nova.virt.hardware [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 756.959030] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36cdb4a0-7289-4745-8d17-0c4b700686ff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.967023] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-028fe769-03d8-459f-afc3-9a1f1f5c5ac7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.981337] env[63293]: ERROR nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 861f3155-fa0f-452c-8005-af67c9fae028, please check neutron logs for more information. [ 756.981337] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Traceback (most recent call last): [ 756.981337] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 756.981337] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] yield resources [ 756.981337] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.981337] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] self.driver.spawn(context, instance, image_meta, [ 756.981337] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 756.981337] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.981337] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.981337] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] vm_ref = self.build_virtual_machine(instance, [ 756.981337] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] for vif in network_info: [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] return self._sync_wrapper(fn, *args, **kwargs) [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] self.wait() [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] self[:] = self._gt.wait() [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] return self._exit_event.wait() [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 756.981658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] current.throw(*self._exc) [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] result = function(*args, **kwargs) [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] return func(*args, **kwargs) [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] raise e [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] nwinfo = self.network_api.allocate_for_instance( [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] created_port_ids = self._update_ports_for_instance( [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] with excutils.save_and_reraise_exception(): [ 756.981955] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.982262] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] self.force_reraise() [ 756.982262] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.982262] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] raise self.value [ 756.982262] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.982262] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] updated_port = self._update_port( [ 756.982262] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.982262] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] _ensure_no_port_binding_failure(port) [ 756.982262] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.982262] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] raise exception.PortBindingFailed(port_id=port['id']) [ 756.982262] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] nova.exception.PortBindingFailed: Binding failed for port 861f3155-fa0f-452c-8005-af67c9fae028, please check neutron logs for more information. [ 756.982262] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] [ 756.982262] env[63293]: INFO nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Terminating instance [ 756.983569] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "refresh_cache-0df77939-d60c-4ccb-98f7-2325062ec76f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.983724] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired lock "refresh_cache-0df77939-d60c-4ccb-98f7-2325062ec76f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.983883] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.370011] env[63293]: DEBUG oslo_vmware.api [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327492, 'name': PowerOnVM_Task, 'duration_secs': 0.43184} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.370275] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 757.370472] env[63293]: INFO nova.compute.manager [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Took 5.33 seconds to spawn the instance on the hypervisor. [ 757.370650] env[63293]: DEBUG nova.compute.manager [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 757.371408] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b842e99-7b72-46b6-945b-b2ec32f179fe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.391449] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.391869] env[63293]: DEBUG nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.394038] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.214s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.395416] env[63293]: INFO nova.compute.claims [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.405096] env[63293]: INFO nova.compute.manager [-] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Took 1.02 seconds to deallocate network for instance. [ 757.406871] env[63293]: DEBUG nova.compute.claims [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 757.407045] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.502609] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.590027] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.887932] env[63293]: INFO nova.compute.manager [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Took 24.98 seconds to build instance. [ 757.899246] env[63293]: DEBUG nova.compute.utils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 757.902748] env[63293]: DEBUG nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 757.902930] env[63293]: DEBUG nova.network.neutron [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 757.953266] env[63293]: DEBUG nova.policy [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1750c074f3bf4bf890b98ec04646e605', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa1d213cb9504f5e954f29217ed447de', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.091999] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Releasing lock "refresh_cache-0df77939-d60c-4ccb-98f7-2325062ec76f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.092669] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 758.092975] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 758.093430] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f8c4860b-d8be-4d2d-9867-6e23f4d367d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.106749] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e7b2834-51c0-46a1-8829-e984039e25ae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.143387] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0df77939-d60c-4ccb-98f7-2325062ec76f could not be found. [ 758.143563] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 758.144193] env[63293]: INFO nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 758.144326] env[63293]: DEBUG oslo.service.loopingcall [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.144707] env[63293]: DEBUG nova.compute.manager [-] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.144904] env[63293]: DEBUG nova.network.neutron [-] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.163556] env[63293]: DEBUG nova.network.neutron [-] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.390050] env[63293]: DEBUG nova.compute.manager [req-7f33d3ae-6bbf-462b-80ad-6e0ecceae64a req-1eda1a56-2cea-4373-a392-ae30f4f58e2e service nova] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Received event network-changed-861f3155-fa0f-452c-8005-af67c9fae028 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.390258] env[63293]: DEBUG nova.compute.manager [req-7f33d3ae-6bbf-462b-80ad-6e0ecceae64a req-1eda1a56-2cea-4373-a392-ae30f4f58e2e service nova] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Refreshing instance network info cache due to event network-changed-861f3155-fa0f-452c-8005-af67c9fae028. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 758.390479] env[63293]: DEBUG oslo_concurrency.lockutils [req-7f33d3ae-6bbf-462b-80ad-6e0ecceae64a req-1eda1a56-2cea-4373-a392-ae30f4f58e2e service nova] Acquiring lock "refresh_cache-0df77939-d60c-4ccb-98f7-2325062ec76f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.390641] env[63293]: DEBUG oslo_concurrency.lockutils [req-7f33d3ae-6bbf-462b-80ad-6e0ecceae64a req-1eda1a56-2cea-4373-a392-ae30f4f58e2e service nova] Acquired lock "refresh_cache-0df77939-d60c-4ccb-98f7-2325062ec76f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.391382] env[63293]: DEBUG nova.network.neutron [req-7f33d3ae-6bbf-462b-80ad-6e0ecceae64a req-1eda1a56-2cea-4373-a392-ae30f4f58e2e service nova] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Refreshing network info cache for port 861f3155-fa0f-452c-8005-af67c9fae028 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 758.392623] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d2247997-70c0-4dc5-a7d8-9c63ff7a1496 tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "b5334832-25dc-4b03-9689-388a1382200a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.223s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.410530] env[63293]: DEBUG nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.507091] env[63293]: DEBUG nova.network.neutron [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Successfully created port: f1003952-8806-4d1c-9b8d-007aa7ea5792 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.666582] env[63293]: DEBUG nova.network.neutron [-] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.820416] env[63293]: INFO nova.compute.manager [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Rebuilding instance [ 758.831733] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e710395-18fd-478b-bdc4-6cb665c65320 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.839847] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5c9e98-ddde-4b41-af1a-a5180bbe39fb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.878477] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d328989-0fd8-453d-ba82-51bcb70de4e6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.886308] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07941d12-8220-4676-8de7-d75f1e31f237 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.902224] env[63293]: DEBUG nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 758.907019] env[63293]: DEBUG nova.compute.provider_tree [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.912883] env[63293]: DEBUG nova.compute.manager [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 758.913967] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145fc813-a008-4824-9d3b-e9b881b2aef6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.987800] env[63293]: DEBUG nova.network.neutron [req-7f33d3ae-6bbf-462b-80ad-6e0ecceae64a req-1eda1a56-2cea-4373-a392-ae30f4f58e2e service nova] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.102886] env[63293]: DEBUG nova.network.neutron [req-7f33d3ae-6bbf-462b-80ad-6e0ecceae64a req-1eda1a56-2cea-4373-a392-ae30f4f58e2e service nova] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.172114] env[63293]: INFO nova.compute.manager [-] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Took 1.03 seconds to deallocate network for instance. [ 759.174597] env[63293]: DEBUG nova.compute.claims [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 759.174778] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.412528] env[63293]: DEBUG nova.scheduler.client.report [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.424592] env[63293]: DEBUG nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.429736] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 759.430411] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5bbfc994-cec3-4118-9830-1a11d26d4b91 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.436938] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.441133] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 759.441133] env[63293]: value = "task-1327493" [ 759.441133] env[63293]: _type = "Task" [ 759.441133] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.450698] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327493, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.455129] env[63293]: DEBUG nova.virt.hardware [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.455129] env[63293]: DEBUG nova.virt.hardware [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.455129] env[63293]: DEBUG nova.virt.hardware [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.455355] env[63293]: DEBUG nova.virt.hardware [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.455355] env[63293]: DEBUG nova.virt.hardware [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.455355] env[63293]: DEBUG nova.virt.hardware [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.455355] env[63293]: DEBUG nova.virt.hardware [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.455355] env[63293]: DEBUG nova.virt.hardware [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.455498] env[63293]: DEBUG nova.virt.hardware [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.455498] env[63293]: DEBUG nova.virt.hardware [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.455715] env[63293]: DEBUG nova.virt.hardware [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.456898] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f30448d-929a-4112-ac7a-1ad34cee2052 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.465495] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6ebef6-b69c-4a4f-be0d-bfc10314ad4a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.604284] env[63293]: DEBUG oslo_concurrency.lockutils [req-7f33d3ae-6bbf-462b-80ad-6e0ecceae64a req-1eda1a56-2cea-4373-a392-ae30f4f58e2e service nova] Releasing lock "refresh_cache-0df77939-d60c-4ccb-98f7-2325062ec76f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.604522] env[63293]: DEBUG nova.compute.manager [req-7f33d3ae-6bbf-462b-80ad-6e0ecceae64a req-1eda1a56-2cea-4373-a392-ae30f4f58e2e service nova] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Received event network-vif-deleted-861f3155-fa0f-452c-8005-af67c9fae028 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.702287] env[63293]: ERROR nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f1003952-8806-4d1c-9b8d-007aa7ea5792, please check neutron logs for more information. [ 759.702287] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 759.702287] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.702287] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 759.702287] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.702287] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 759.702287] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.702287] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 759.702287] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.702287] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 759.702287] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.702287] env[63293]: ERROR nova.compute.manager raise self.value [ 759.702287] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.702287] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 759.702287] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.702287] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 759.702677] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.702677] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 759.702677] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f1003952-8806-4d1c-9b8d-007aa7ea5792, please check neutron logs for more information. [ 759.702677] env[63293]: ERROR nova.compute.manager [ 759.702677] env[63293]: Traceback (most recent call last): [ 759.702677] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 759.702677] env[63293]: listener.cb(fileno) [ 759.702677] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.702677] env[63293]: result = function(*args, **kwargs) [ 759.702677] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 759.702677] env[63293]: return func(*args, **kwargs) [ 759.702677] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.702677] env[63293]: raise e [ 759.702677] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.702677] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 759.702677] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.702677] env[63293]: created_port_ids = self._update_ports_for_instance( [ 759.702677] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.702677] env[63293]: with excutils.save_and_reraise_exception(): [ 759.702677] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.702677] env[63293]: self.force_reraise() [ 759.702677] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.702677] env[63293]: raise self.value [ 759.702677] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.702677] env[63293]: updated_port = self._update_port( [ 759.702677] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.702677] env[63293]: _ensure_no_port_binding_failure(port) [ 759.702677] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.702677] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 759.703255] env[63293]: nova.exception.PortBindingFailed: Binding failed for port f1003952-8806-4d1c-9b8d-007aa7ea5792, please check neutron logs for more information. [ 759.703255] env[63293]: Removing descriptor: 15 [ 759.703255] env[63293]: ERROR nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f1003952-8806-4d1c-9b8d-007aa7ea5792, please check neutron logs for more information. [ 759.703255] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Traceback (most recent call last): [ 759.703255] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 759.703255] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] yield resources [ 759.703255] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 759.703255] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] self.driver.spawn(context, instance, image_meta, [ 759.703255] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 759.703255] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.703255] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.703255] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] vm_ref = self.build_virtual_machine(instance, [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] for vif in network_info: [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] return self._sync_wrapper(fn, *args, **kwargs) [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] self.wait() [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] self[:] = self._gt.wait() [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] return self._exit_event.wait() [ 759.703500] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] result = hub.switch() [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] return self.greenlet.switch() [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] result = function(*args, **kwargs) [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] return func(*args, **kwargs) [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] raise e [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] nwinfo = self.network_api.allocate_for_instance( [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.703754] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] created_port_ids = self._update_ports_for_instance( [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] with excutils.save_and_reraise_exception(): [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] self.force_reraise() [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] raise self.value [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] updated_port = self._update_port( [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] _ensure_no_port_binding_failure(port) [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.704009] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] raise exception.PortBindingFailed(port_id=port['id']) [ 759.704263] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] nova.exception.PortBindingFailed: Binding failed for port f1003952-8806-4d1c-9b8d-007aa7ea5792, please check neutron logs for more information. [ 759.704263] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] [ 759.704263] env[63293]: INFO nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Terminating instance [ 759.705859] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Acquiring lock "refresh_cache-b1b1ebe0-187b-4e23-9433-7dff420e3de3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.706036] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Acquired lock "refresh_cache-b1b1ebe0-187b-4e23-9433-7dff420e3de3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.706203] env[63293]: DEBUG nova.network.neutron [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.919075] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.919075] env[63293]: DEBUG nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 759.922410] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.515s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.949098] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327493, 'name': PowerOffVM_Task, 'duration_secs': 0.189863} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.949361] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 759.949772] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.950635] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8c1981-aea5-4dad-930a-fa97da314399 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.957849] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 759.958755] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28ad7964-31a3-494b-920e-eac8564304f0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.985741] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 759.985954] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 759.986147] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Deleting the datastore file [datastore1] b5334832-25dc-4b03-9689-388a1382200a {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 759.986405] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f32e5c9-a35f-4bb1-ab1a-f136f6dba2e7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.993109] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 759.993109] env[63293]: value = "task-1327495" [ 759.993109] env[63293]: _type = "Task" [ 759.993109] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.001099] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327495, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.225419] env[63293]: DEBUG nova.network.neutron [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.299430] env[63293]: DEBUG nova.network.neutron [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.413503] env[63293]: DEBUG nova.compute.manager [req-fce0ba73-92ef-4348-87e5-982dfca9fff6 req-b97869f0-97d7-45d2-9aa3-e3e746e6fe20 service nova] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Received event network-changed-f1003952-8806-4d1c-9b8d-007aa7ea5792 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.413695] env[63293]: DEBUG nova.compute.manager [req-fce0ba73-92ef-4348-87e5-982dfca9fff6 req-b97869f0-97d7-45d2-9aa3-e3e746e6fe20 service nova] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Refreshing instance network info cache due to event network-changed-f1003952-8806-4d1c-9b8d-007aa7ea5792. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 760.413885] env[63293]: DEBUG oslo_concurrency.lockutils [req-fce0ba73-92ef-4348-87e5-982dfca9fff6 req-b97869f0-97d7-45d2-9aa3-e3e746e6fe20 service nova] Acquiring lock "refresh_cache-b1b1ebe0-187b-4e23-9433-7dff420e3de3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.426942] env[63293]: DEBUG nova.compute.utils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.430800] env[63293]: DEBUG nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 760.430976] env[63293]: DEBUG nova.network.neutron [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 760.474045] env[63293]: DEBUG nova.policy [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c78ec5284164f19bf27698c667ac630', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff9b6966a9a5406da28f997b19e58139', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 760.503338] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327495, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.789436] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e980cfc8-de0e-4d1d-a6d2-a9c022327aad {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.799165] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d27d995-6973-4569-b6e2-8700b7a1ad37 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.802788] env[63293]: DEBUG nova.network.neutron [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Successfully created port: 7707013f-9522-41aa-b0a5-c3b6d07a8bf3 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.804621] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Releasing lock "refresh_cache-b1b1ebe0-187b-4e23-9433-7dff420e3de3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.804994] env[63293]: DEBUG nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 760.805201] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 760.829771] env[63293]: DEBUG oslo_concurrency.lockutils [req-fce0ba73-92ef-4348-87e5-982dfca9fff6 req-b97869f0-97d7-45d2-9aa3-e3e746e6fe20 service nova] Acquired lock "refresh_cache-b1b1ebe0-187b-4e23-9433-7dff420e3de3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.829984] env[63293]: DEBUG nova.network.neutron [req-fce0ba73-92ef-4348-87e5-982dfca9fff6 req-b97869f0-97d7-45d2-9aa3-e3e746e6fe20 service nova] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Refreshing network info cache for port f1003952-8806-4d1c-9b8d-007aa7ea5792 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 760.831076] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db085d35-d0fb-4a52-aa3b-579b32908fd8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.834850] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba7f2a1-a76e-4181-95cc-d2c15edc58bb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.842822] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb93bbc4-8c41-4665-98a2-2098c48cf2c6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.850686] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7ed6e1-b0ff-4067-a73f-69bd31f4711b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.875131] env[63293]: DEBUG nova.compute.provider_tree [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.882201] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b1b1ebe0-187b-4e23-9433-7dff420e3de3 could not be found. [ 760.882489] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 760.882886] env[63293]: INFO nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Took 0.08 seconds to destroy the instance on the hypervisor. [ 760.883054] env[63293]: DEBUG oslo.service.loopingcall [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.883584] env[63293]: DEBUG nova.compute.manager [-] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.886323] env[63293]: DEBUG nova.network.neutron [-] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.914590] env[63293]: DEBUG nova.network.neutron [-] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.934340] env[63293]: DEBUG nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 761.004074] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327495, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.351925] env[63293]: DEBUG nova.network.neutron [req-fce0ba73-92ef-4348-87e5-982dfca9fff6 req-b97869f0-97d7-45d2-9aa3-e3e746e6fe20 service nova] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.387020] env[63293]: DEBUG nova.scheduler.client.report [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.417339] env[63293]: DEBUG nova.network.neutron [-] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.439612] env[63293]: INFO nova.virt.block_device [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Booting with volume 447ac8dc-ca34-484c-b22a-4040db3a664d at /dev/sda [ 761.491126] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf309580-ac51-4459-8988-f2e3108e4bd5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.500611] env[63293]: DEBUG nova.network.neutron [req-fce0ba73-92ef-4348-87e5-982dfca9fff6 req-b97869f0-97d7-45d2-9aa3-e3e746e6fe20 service nova] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.505148] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea682b5-630b-464c-b10f-645eccbc66ff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.522349] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327495, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.508292} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.523117] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 761.523461] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 761.524074] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 761.537195] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9766bc7-e59a-4cc5-a250-44e75ea2cba8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.544749] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70821b28-cdd7-4af7-8a32-754ddc24e8ae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.569094] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b87666e-323b-4297-acf1-b7f0b55d180b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.576184] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d701dc0-6809-4601-836e-cb4635b72185 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.591292] env[63293]: DEBUG nova.virt.block_device [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Updating existing volume attachment record: f60484ac-8258-4d06-a8a8-c7980972c187 {{(pid=63293) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 761.805998] env[63293]: ERROR nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7707013f-9522-41aa-b0a5-c3b6d07a8bf3, please check neutron logs for more information. [ 761.805998] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 761.805998] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.805998] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 761.805998] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.805998] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 761.805998] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.805998] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 761.805998] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.805998] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 761.805998] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.805998] env[63293]: ERROR nova.compute.manager raise self.value [ 761.805998] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.805998] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 761.805998] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.805998] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 761.806602] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.806602] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 761.806602] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7707013f-9522-41aa-b0a5-c3b6d07a8bf3, please check neutron logs for more information. [ 761.806602] env[63293]: ERROR nova.compute.manager [ 761.806602] env[63293]: Traceback (most recent call last): [ 761.806602] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 761.806602] env[63293]: listener.cb(fileno) [ 761.806602] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.806602] env[63293]: result = function(*args, **kwargs) [ 761.806602] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.806602] env[63293]: return func(*args, **kwargs) [ 761.806602] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.806602] env[63293]: raise e [ 761.806602] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.806602] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 761.806602] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.806602] env[63293]: created_port_ids = self._update_ports_for_instance( [ 761.806602] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.806602] env[63293]: with excutils.save_and_reraise_exception(): [ 761.806602] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.806602] env[63293]: self.force_reraise() [ 761.806602] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.806602] env[63293]: raise self.value [ 761.806602] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.806602] env[63293]: updated_port = self._update_port( [ 761.806602] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.806602] env[63293]: _ensure_no_port_binding_failure(port) [ 761.806602] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.806602] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 761.807412] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 7707013f-9522-41aa-b0a5-c3b6d07a8bf3, please check neutron logs for more information. [ 761.807412] env[63293]: Removing descriptor: 15 [ 761.893065] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.971s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.893823] env[63293]: ERROR nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a0334b0c-e404-47dd-a833-3dcee4aeb4ba, please check neutron logs for more information. [ 761.893823] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Traceback (most recent call last): [ 761.893823] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.893823] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] self.driver.spawn(context, instance, image_meta, [ 761.893823] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 761.893823] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.893823] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.893823] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] vm_ref = self.build_virtual_machine(instance, [ 761.893823] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.893823] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.893823] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] for vif in network_info: [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] return self._sync_wrapper(fn, *args, **kwargs) [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] self.wait() [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] self[:] = self._gt.wait() [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] return self._exit_event.wait() [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] result = hub.switch() [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.894131] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] return self.greenlet.switch() [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] result = function(*args, **kwargs) [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] return func(*args, **kwargs) [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] raise e [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] nwinfo = self.network_api.allocate_for_instance( [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] created_port_ids = self._update_ports_for_instance( [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] with excutils.save_and_reraise_exception(): [ 761.894460] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.894787] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] self.force_reraise() [ 761.894787] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.894787] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] raise self.value [ 761.894787] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.894787] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] updated_port = self._update_port( [ 761.894787] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.894787] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] _ensure_no_port_binding_failure(port) [ 761.894787] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.894787] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] raise exception.PortBindingFailed(port_id=port['id']) [ 761.894787] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] nova.exception.PortBindingFailed: Binding failed for port a0334b0c-e404-47dd-a833-3dcee4aeb4ba, please check neutron logs for more information. [ 761.894787] env[63293]: ERROR nova.compute.manager [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] [ 761.895146] env[63293]: DEBUG nova.compute.utils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Binding failed for port a0334b0c-e404-47dd-a833-3dcee4aeb4ba, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 761.895921] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.025s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.896157] env[63293]: DEBUG nova.objects.instance [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lazy-loading 'resources' on Instance uuid 0aade4da-882a-4342-aefb-07eb306b17ff {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 761.898086] env[63293]: DEBUG nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Build of instance db04bc40-1106-46dd-ad9a-f9b11a6c8774 was re-scheduled: Binding failed for port a0334b0c-e404-47dd-a833-3dcee4aeb4ba, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 761.898605] env[63293]: DEBUG nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 761.898806] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "refresh_cache-db04bc40-1106-46dd-ad9a-f9b11a6c8774" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.898968] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquired lock "refresh_cache-db04bc40-1106-46dd-ad9a-f9b11a6c8774" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.899114] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.921175] env[63293]: INFO nova.compute.manager [-] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Took 1.04 seconds to deallocate network for instance. [ 761.922616] env[63293]: DEBUG nova.compute.claims [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 761.922827] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.018918] env[63293]: DEBUG oslo_concurrency.lockutils [req-fce0ba73-92ef-4348-87e5-982dfca9fff6 req-b97869f0-97d7-45d2-9aa3-e3e746e6fe20 service nova] Releasing lock "refresh_cache-b1b1ebe0-187b-4e23-9433-7dff420e3de3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.019210] env[63293]: DEBUG nova.compute.manager [req-fce0ba73-92ef-4348-87e5-982dfca9fff6 req-b97869f0-97d7-45d2-9aa3-e3e746e6fe20 service nova] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Received event network-vif-deleted-f1003952-8806-4d1c-9b8d-007aa7ea5792 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.419300] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.438090] env[63293]: DEBUG nova.compute.manager [req-044ffeab-37af-4171-8e9f-0adc56cdb8d3 req-e5cfe3e3-0985-45a8-9900-ccd72a61a0ce service nova] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Received event network-changed-7707013f-9522-41aa-b0a5-c3b6d07a8bf3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.438305] env[63293]: DEBUG nova.compute.manager [req-044ffeab-37af-4171-8e9f-0adc56cdb8d3 req-e5cfe3e3-0985-45a8-9900-ccd72a61a0ce service nova] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Refreshing instance network info cache due to event network-changed-7707013f-9522-41aa-b0a5-c3b6d07a8bf3. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 762.438508] env[63293]: DEBUG oslo_concurrency.lockutils [req-044ffeab-37af-4171-8e9f-0adc56cdb8d3 req-e5cfe3e3-0985-45a8-9900-ccd72a61a0ce service nova] Acquiring lock "refresh_cache-8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.438657] env[63293]: DEBUG oslo_concurrency.lockutils [req-044ffeab-37af-4171-8e9f-0adc56cdb8d3 req-e5cfe3e3-0985-45a8-9900-ccd72a61a0ce service nova] Acquired lock "refresh_cache-8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.438832] env[63293]: DEBUG nova.network.neutron [req-044ffeab-37af-4171-8e9f-0adc56cdb8d3 req-e5cfe3e3-0985-45a8-9900-ccd72a61a0ce service nova] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Refreshing network info cache for port 7707013f-9522-41aa-b0a5-c3b6d07a8bf3 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 762.501922] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.563840] env[63293]: DEBUG nova.virt.hardware [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.564122] env[63293]: DEBUG nova.virt.hardware [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.564307] env[63293]: DEBUG nova.virt.hardware [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.564526] env[63293]: DEBUG nova.virt.hardware [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.564697] env[63293]: DEBUG nova.virt.hardware [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.564866] env[63293]: DEBUG nova.virt.hardware [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.565150] env[63293]: DEBUG nova.virt.hardware [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.565411] env[63293]: DEBUG nova.virt.hardware [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.565569] env[63293]: DEBUG nova.virt.hardware [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.565755] env[63293]: DEBUG nova.virt.hardware [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.566149] env[63293]: DEBUG nova.virt.hardware [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.566874] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d33bc0-6d86-4278-9777-f03ae6455577 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.577952] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3fe9c1-590b-4128-bcae-3dac5a537e48 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.593846] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 762.599364] env[63293]: DEBUG oslo.service.loopingcall [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.601965] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5334832-25dc-4b03-9689-388a1382200a] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 762.602720] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ee28af4-e84b-43b9-bfaf-c30d65d98350 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.619928] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 762.619928] env[63293]: value = "task-1327496" [ 762.619928] env[63293]: _type = "Task" [ 762.619928] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.627382] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327496, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.776435] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45e6899-de4f-4351-bbe5-2c3783cca05a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.783939] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb635ae-76e7-43df-a9cf-f1be15c0215f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.815575] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d871d66a-ab3b-46cc-904b-6811456cedfc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.823110] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c779701-48b5-4882-a64a-595149ad75ab {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.839423] env[63293]: DEBUG nova.compute.provider_tree [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.959471] env[63293]: DEBUG nova.network.neutron [req-044ffeab-37af-4171-8e9f-0adc56cdb8d3 req-e5cfe3e3-0985-45a8-9900-ccd72a61a0ce service nova] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.005567] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Releasing lock "refresh_cache-db04bc40-1106-46dd-ad9a-f9b11a6c8774" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.005698] env[63293]: DEBUG nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 763.005835] env[63293]: DEBUG nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.006049] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.029091] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.067400] env[63293]: DEBUG nova.network.neutron [req-044ffeab-37af-4171-8e9f-0adc56cdb8d3 req-e5cfe3e3-0985-45a8-9900-ccd72a61a0ce service nova] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.129809] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327496, 'name': CreateVM_Task, 'duration_secs': 0.256081} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.129976] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5334832-25dc-4b03-9689-388a1382200a] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 763.130385] env[63293]: DEBUG oslo_concurrency.lockutils [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.130544] env[63293]: DEBUG oslo_concurrency.lockutils [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.130895] env[63293]: DEBUG oslo_concurrency.lockutils [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 763.131214] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b4d33e6-e1e4-4ac0-b97c-4c6d1b4ede07 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.136443] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 763.136443] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a40f76-00fc-4631-b92d-ddef1b05bb9b" [ 763.136443] env[63293]: _type = "Task" [ 763.136443] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.147301] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a40f76-00fc-4631-b92d-ddef1b05bb9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.343182] env[63293]: DEBUG nova.scheduler.client.report [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.531495] env[63293]: DEBUG nova.network.neutron [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.569624] env[63293]: DEBUG oslo_concurrency.lockutils [req-044ffeab-37af-4171-8e9f-0adc56cdb8d3 req-e5cfe3e3-0985-45a8-9900-ccd72a61a0ce service nova] Releasing lock "refresh_cache-8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.569913] env[63293]: DEBUG nova.compute.manager [req-044ffeab-37af-4171-8e9f-0adc56cdb8d3 req-e5cfe3e3-0985-45a8-9900-ccd72a61a0ce service nova] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Received event network-vif-deleted-7707013f-9522-41aa-b0a5-c3b6d07a8bf3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 763.647785] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a40f76-00fc-4631-b92d-ddef1b05bb9b, 'name': SearchDatastore_Task, 'duration_secs': 0.010271} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.648111] env[63293]: DEBUG oslo_concurrency.lockutils [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.648346] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 763.648619] env[63293]: DEBUG oslo_concurrency.lockutils [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.648782] env[63293]: DEBUG oslo_concurrency.lockutils [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.648965] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 763.649239] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff585cc5-3518-43e0-9c7a-ebc57c28f2a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.656598] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 763.656813] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 763.657479] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-505d7ccc-58b9-45da-9d86-81b5e01d3cd4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.662436] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 763.662436] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5293e927-9e59-2d7b-73fc-dd558748a397" [ 763.662436] env[63293]: _type = "Task" [ 763.662436] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.670486] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5293e927-9e59-2d7b-73fc-dd558748a397, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.698022] env[63293]: DEBUG nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 763.698616] env[63293]: DEBUG nova.virt.hardware [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 763.698861] env[63293]: DEBUG nova.virt.hardware [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 763.698990] env[63293]: DEBUG nova.virt.hardware [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 763.699185] env[63293]: DEBUG nova.virt.hardware [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 763.699331] env[63293]: DEBUG nova.virt.hardware [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 763.699475] env[63293]: DEBUG nova.virt.hardware [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 763.699701] env[63293]: DEBUG nova.virt.hardware [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 763.699866] env[63293]: DEBUG nova.virt.hardware [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 763.700045] env[63293]: DEBUG nova.virt.hardware [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 763.700213] env[63293]: DEBUG nova.virt.hardware [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 763.700385] env[63293]: DEBUG nova.virt.hardware [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.701317] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3e694f-0631-490d-9744-b311608ce8cd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.709258] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a6ef95-30e4-40e3-b5fe-750f5b5454c1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.723928] env[63293]: ERROR nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7707013f-9522-41aa-b0a5-c3b6d07a8bf3, please check neutron logs for more information. [ 763.723928] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Traceback (most recent call last): [ 763.723928] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 763.723928] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] yield resources [ 763.723928] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 763.723928] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] self.driver.spawn(context, instance, image_meta, [ 763.723928] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 763.723928] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 763.723928] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 763.723928] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] vm_ref = self.build_virtual_machine(instance, [ 763.723928] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] vif_infos = vmwarevif.get_vif_info(self._session, [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] for vif in network_info: [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] return self._sync_wrapper(fn, *args, **kwargs) [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] self.wait() [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] self[:] = self._gt.wait() [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] return self._exit_event.wait() [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 763.724357] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] current.throw(*self._exc) [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] result = function(*args, **kwargs) [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] return func(*args, **kwargs) [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] raise e [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] nwinfo = self.network_api.allocate_for_instance( [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] created_port_ids = self._update_ports_for_instance( [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] with excutils.save_and_reraise_exception(): [ 763.724663] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.724992] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] self.force_reraise() [ 763.724992] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.724992] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] raise self.value [ 763.724992] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.724992] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] updated_port = self._update_port( [ 763.724992] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.724992] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] _ensure_no_port_binding_failure(port) [ 763.724992] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.724992] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] raise exception.PortBindingFailed(port_id=port['id']) [ 763.724992] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] nova.exception.PortBindingFailed: Binding failed for port 7707013f-9522-41aa-b0a5-c3b6d07a8bf3, please check neutron logs for more information. [ 763.724992] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] [ 763.724992] env[63293]: INFO nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Terminating instance [ 763.726222] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Acquiring lock "refresh_cache-8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.726377] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Acquired lock "refresh_cache-8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.726539] env[63293]: DEBUG nova.network.neutron [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.848345] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.952s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.850838] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.337s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.866558] env[63293]: INFO nova.scheduler.client.report [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Deleted allocations for instance 0aade4da-882a-4342-aefb-07eb306b17ff [ 764.034860] env[63293]: INFO nova.compute.manager [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: db04bc40-1106-46dd-ad9a-f9b11a6c8774] Took 1.03 seconds to deallocate network for instance. [ 764.173405] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5293e927-9e59-2d7b-73fc-dd558748a397, 'name': SearchDatastore_Task, 'duration_secs': 0.009311} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.174198] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8a58e9a-887f-4bac-9638-efc4c32e6a2e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.179480] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 764.179480] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524e110e-e9f8-a7c8-d93b-133395da87ea" [ 764.179480] env[63293]: _type = "Task" [ 764.179480] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.186649] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524e110e-e9f8-a7c8-d93b-133395da87ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.244096] env[63293]: DEBUG nova.network.neutron [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.327945] env[63293]: DEBUG nova.network.neutron [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.374111] env[63293]: DEBUG oslo_concurrency.lockutils [None req-607ca090-f09d-4a24-9044-860d58acac49 tempest-ServersAaction247Test-1620350231 tempest-ServersAaction247Test-1620350231-project-member] Lock "0aade4da-882a-4342-aefb-07eb306b17ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.293s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.672798] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db37dcab-73b1-4d4b-8405-c4e79a60c69d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.683636] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f648a5dc-8da8-4b96-83b9-f8d7475def98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.691572] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524e110e-e9f8-a7c8-d93b-133395da87ea, 'name': SearchDatastore_Task, 'duration_secs': 0.009068} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.716622] env[63293]: DEBUG oslo_concurrency.lockutils [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.716811] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] b5334832-25dc-4b03-9689-388a1382200a/b5334832-25dc-4b03-9689-388a1382200a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 764.717291] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d0a1246-49b8-42ce-87cf-47a0dee72be6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.719796] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf01ffff-8eaa-4032-b129-e282cc3ddc6a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.728716] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d357807-76f8-4264-98fb-f3ef3a2004cc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.732851] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 764.732851] env[63293]: value = "task-1327497" [ 764.732851] env[63293]: _type = "Task" [ 764.732851] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.745384] env[63293]: DEBUG nova.compute.provider_tree [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.755011] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327497, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.830683] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Releasing lock "refresh_cache-8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.831364] env[63293]: DEBUG nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 764.831721] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c01d2d7-2df5-4ebf-9936-11cb2df45744 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.841216] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fec1391-9ee2-446b-b84b-7f54885f4b97 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.863270] env[63293]: WARNING nova.virt.vmwareapi.driver [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a could not be found. [ 764.863451] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 764.863735] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-af5ace9c-dfdc-4d33-9975-dce14c9c77fa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.871327] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d86edc-6cfd-441d-bf2e-cff09c384c79 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.894188] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a could not be found. [ 764.894333] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 764.894518] env[63293]: INFO nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Took 0.06 seconds to destroy the instance on the hypervisor. [ 764.894788] env[63293]: DEBUG oslo.service.loopingcall [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.895096] env[63293]: DEBUG nova.compute.manager [-] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.895196] env[63293]: DEBUG nova.network.neutron [-] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 764.912023] env[63293]: DEBUG nova.network.neutron [-] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.071422] env[63293]: INFO nova.scheduler.client.report [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Deleted allocations for instance db04bc40-1106-46dd-ad9a-f9b11a6c8774 [ 765.244026] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327497, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487645} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.244026] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] b5334832-25dc-4b03-9689-388a1382200a/b5334832-25dc-4b03-9689-388a1382200a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 765.244026] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 765.244258] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ad45bcc-016b-4f58-a636-45e4c342d2c6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.248924] env[63293]: DEBUG nova.scheduler.client.report [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.253071] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 765.253071] env[63293]: value = "task-1327498" [ 765.253071] env[63293]: _type = "Task" [ 765.253071] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.263437] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327498, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.414511] env[63293]: DEBUG nova.network.neutron [-] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.585214] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c5842aae-8df0-43fb-b56a-cbd3749e4582 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "db04bc40-1106-46dd-ad9a-f9b11a6c8774" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 178.677s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.755053] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.904s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.755642] env[63293]: ERROR nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 37e8a179-ff61-48b6-8d49-44c09262ea80, please check neutron logs for more information. [ 765.755642] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Traceback (most recent call last): [ 765.755642] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.755642] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] self.driver.spawn(context, instance, image_meta, [ 765.755642] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 765.755642] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.755642] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.755642] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] vm_ref = self.build_virtual_machine(instance, [ 765.755642] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.755642] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.755642] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] for vif in network_info: [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] return self._sync_wrapper(fn, *args, **kwargs) [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] self.wait() [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] self[:] = self._gt.wait() [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] return self._exit_event.wait() [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] result = hub.switch() [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 765.755942] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] return self.greenlet.switch() [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] result = function(*args, **kwargs) [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] return func(*args, **kwargs) [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] raise e [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] nwinfo = self.network_api.allocate_for_instance( [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] created_port_ids = self._update_ports_for_instance( [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] with excutils.save_and_reraise_exception(): [ 765.756357] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.756689] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] self.force_reraise() [ 765.756689] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.756689] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] raise self.value [ 765.756689] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.756689] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] updated_port = self._update_port( [ 765.756689] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.756689] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] _ensure_no_port_binding_failure(port) [ 765.756689] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.756689] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] raise exception.PortBindingFailed(port_id=port['id']) [ 765.756689] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] nova.exception.PortBindingFailed: Binding failed for port 37e8a179-ff61-48b6-8d49-44c09262ea80, please check neutron logs for more information. [ 765.756689] env[63293]: ERROR nova.compute.manager [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] [ 765.756960] env[63293]: DEBUG nova.compute.utils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Binding failed for port 37e8a179-ff61-48b6-8d49-44c09262ea80, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 765.757676] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.115s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.757852] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.757992] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63293) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 765.758262] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.657s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.763997] env[63293]: DEBUG nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Build of instance 263d1329-a81a-4d50-91b6-3927bffe8d4c was re-scheduled: Binding failed for port 37e8a179-ff61-48b6-8d49-44c09262ea80, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 765.764443] env[63293]: DEBUG nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 765.764665] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Acquiring lock "refresh_cache-263d1329-a81a-4d50-91b6-3927bffe8d4c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.764811] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Acquired lock "refresh_cache-263d1329-a81a-4d50-91b6-3927bffe8d4c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.764963] env[63293]: DEBUG nova.network.neutron [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.766633] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a719398a-f05a-4c31-b22f-318acbc88aa5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.773996] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327498, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066201} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.776931] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 765.777772] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa116941-daed-42cb-b237-d2317a7327ca {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.781028] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0443a5-a56c-4717-acf2-019af406f650 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.802161] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb2b7121-3081-459b-b7a3-167f24ec61b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.813631] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] b5334832-25dc-4b03-9689-388a1382200a/b5334832-25dc-4b03-9689-388a1382200a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 765.814221] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-177d022d-96d2-4ee5-a129-b4fb5e959589 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.833561] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dab311a-caed-4705-8de5-a8037cebd1e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.837911] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 765.837911] env[63293]: value = "task-1327499" [ 765.837911] env[63293]: _type = "Task" [ 765.837911] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.867804] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181455MB free_disk=132GB free_vcpus=48 pci_devices=None {{(pid=63293) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 765.867968] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.874317] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.918624] env[63293]: INFO nova.compute.manager [-] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Took 1.02 seconds to deallocate network for instance. [ 766.089750] env[63293]: DEBUG nova.compute.manager [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.287821] env[63293]: DEBUG nova.network.neutron [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.349300] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327499, 'name': ReconfigVM_Task, 'duration_secs': 0.278991} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.351616] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Reconfigured VM instance instance-00000030 to attach disk [datastore2] b5334832-25dc-4b03-9689-388a1382200a/b5334832-25dc-4b03-9689-388a1382200a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 766.352146] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a2388d70-660b-4a99-b06b-b16c3fea9bba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.359469] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 766.359469] env[63293]: value = "task-1327500" [ 766.359469] env[63293]: _type = "Task" [ 766.359469] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.367696] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327500, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.390860] env[63293]: DEBUG nova.network.neutron [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.485720] env[63293]: INFO nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Took 0.57 seconds to detach 1 volumes for instance. [ 766.490129] env[63293]: DEBUG nova.compute.claims [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 766.490345] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.612048] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.619571] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c27594a-92ac-4e75-808c-8d6bd8ff0262 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.626863] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194e06ec-28c6-47d7-9572-73b216cf601d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.658186] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0daa8a-4c40-41cf-8b9b-950af6935afa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.665634] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eab00d2-ed39-47c7-9a03-9fbc892378d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.678759] env[63293]: DEBUG nova.compute.provider_tree [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.871447] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327500, 'name': Rename_Task, 'duration_secs': 0.135628} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.871665] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 766.871906] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc1975f8-51d7-46c8-a259-8faadccb7358 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.877571] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 766.877571] env[63293]: value = "task-1327501" [ 766.877571] env[63293]: _type = "Task" [ 766.877571] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.884899] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327501, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.895952] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Releasing lock "refresh_cache-263d1329-a81a-4d50-91b6-3927bffe8d4c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.895952] env[63293]: DEBUG nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 766.895952] env[63293]: DEBUG nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.896136] env[63293]: DEBUG nova.network.neutron [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.915481] env[63293]: DEBUG nova.network.neutron [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.185957] env[63293]: DEBUG nova.scheduler.client.report [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.388193] env[63293]: DEBUG oslo_vmware.api [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327501, 'name': PowerOnVM_Task, 'duration_secs': 0.434214} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.388500] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 767.388817] env[63293]: DEBUG nova.compute.manager [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.389628] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e82dcc5-0aff-48f8-aa7d-e3663ff87d47 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.418749] env[63293]: DEBUG nova.network.neutron [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.689217] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.689922] env[63293]: ERROR nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0845133e-ee59-4fca-8d15-f336202ab231, please check neutron logs for more information. [ 767.689922] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Traceback (most recent call last): [ 767.689922] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.689922] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] self.driver.spawn(context, instance, image_meta, [ 767.689922] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 767.689922] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.689922] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.689922] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] vm_ref = self.build_virtual_machine(instance, [ 767.689922] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.689922] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.689922] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] for vif in network_info: [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] return self._sync_wrapper(fn, *args, **kwargs) [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] self.wait() [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] self[:] = self._gt.wait() [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] return self._exit_event.wait() [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] result = hub.switch() [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 767.690368] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] return self.greenlet.switch() [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] result = function(*args, **kwargs) [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] return func(*args, **kwargs) [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] raise e [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] nwinfo = self.network_api.allocate_for_instance( [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] created_port_ids = self._update_ports_for_instance( [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] with excutils.save_and_reraise_exception(): [ 767.690738] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.691106] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] self.force_reraise() [ 767.691106] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.691106] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] raise self.value [ 767.691106] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.691106] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] updated_port = self._update_port( [ 767.691106] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.691106] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] _ensure_no_port_binding_failure(port) [ 767.691106] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.691106] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] raise exception.PortBindingFailed(port_id=port['id']) [ 767.691106] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] nova.exception.PortBindingFailed: Binding failed for port 0845133e-ee59-4fca-8d15-f336202ab231, please check neutron logs for more information. [ 767.691106] env[63293]: ERROR nova.compute.manager [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] [ 767.691367] env[63293]: DEBUG nova.compute.utils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Binding failed for port 0845133e-ee59-4fca-8d15-f336202ab231, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 767.691898] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.337s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.694853] env[63293]: DEBUG nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Build of instance fd67940d-9552-42b3-b472-2ce0e7899ea6 was re-scheduled: Binding failed for port 0845133e-ee59-4fca-8d15-f336202ab231, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 767.695376] env[63293]: DEBUG nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 767.695608] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-fd67940d-9552-42b3-b472-2ce0e7899ea6" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.695755] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-fd67940d-9552-42b3-b472-2ce0e7899ea6" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.695985] env[63293]: DEBUG nova.network.neutron [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.909802] env[63293]: DEBUG oslo_concurrency.lockutils [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.922507] env[63293]: INFO nova.compute.manager [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] [instance: 263d1329-a81a-4d50-91b6-3927bffe8d4c] Took 1.03 seconds to deallocate network for instance. [ 768.217922] env[63293]: DEBUG nova.network.neutron [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.324249] env[63293]: DEBUG nova.network.neutron [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.561377] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81152eaa-7084-43e8-aa2b-248ea5fde552 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.569632] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7e7113-0292-4b84-a382-04b386c0d912 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.602427] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0929e5-302e-4a9c-851c-0d0c7c1ed725 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.610264] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739ccc18-6d08-4aad-93a4-c15597072aa2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.623881] env[63293]: DEBUG nova.compute.provider_tree [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.829060] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-fd67940d-9552-42b3-b472-2ce0e7899ea6" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.829980] env[63293]: DEBUG nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 768.829980] env[63293]: DEBUG nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.829980] env[63293]: DEBUG nova.network.neutron [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.845136] env[63293]: DEBUG nova.network.neutron [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.850670] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "f427933b-dd2c-49a9-b401-a647635a1eab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.851848] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "f427933b-dd2c-49a9-b401-a647635a1eab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.954983] env[63293]: INFO nova.scheduler.client.report [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Deleted allocations for instance 263d1329-a81a-4d50-91b6-3927bffe8d4c [ 768.973802] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquiring lock "b5334832-25dc-4b03-9689-388a1382200a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.974033] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "b5334832-25dc-4b03-9689-388a1382200a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.974246] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquiring lock "b5334832-25dc-4b03-9689-388a1382200a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.974428] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "b5334832-25dc-4b03-9689-388a1382200a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.974596] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "b5334832-25dc-4b03-9689-388a1382200a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.976657] env[63293]: INFO nova.compute.manager [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Terminating instance [ 768.978280] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquiring lock "refresh_cache-b5334832-25dc-4b03-9689-388a1382200a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.978432] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquired lock "refresh_cache-b5334832-25dc-4b03-9689-388a1382200a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.978641] env[63293]: DEBUG nova.network.neutron [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.126588] env[63293]: DEBUG nova.scheduler.client.report [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.351699] env[63293]: DEBUG nova.network.neutron [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.463524] env[63293]: DEBUG oslo_concurrency.lockutils [None req-28cba3d3-9efa-4555-aa1b-854e56d22c39 tempest-ServersTestJSON-607300979 tempest-ServersTestJSON-607300979-project-member] Lock "263d1329-a81a-4d50-91b6-3927bffe8d4c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.458s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.496024] env[63293]: DEBUG nova.network.neutron [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.569571] env[63293]: DEBUG nova.network.neutron [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.631648] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.940s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.632342] env[63293]: ERROR nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dbed6212-8b1d-4e39-bcd7-b0cf04b60660, please check neutron logs for more information. [ 769.632342] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Traceback (most recent call last): [ 769.632342] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.632342] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] self.driver.spawn(context, instance, image_meta, [ 769.632342] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 769.632342] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.632342] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.632342] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] vm_ref = self.build_virtual_machine(instance, [ 769.632342] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.632342] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.632342] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] for vif in network_info: [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] return self._sync_wrapper(fn, *args, **kwargs) [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] self.wait() [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] self[:] = self._gt.wait() [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] return self._exit_event.wait() [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] current.throw(*self._exc) [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.632661] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] result = function(*args, **kwargs) [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] return func(*args, **kwargs) [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] raise e [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] nwinfo = self.network_api.allocate_for_instance( [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] created_port_ids = self._update_ports_for_instance( [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] with excutils.save_and_reraise_exception(): [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] self.force_reraise() [ 769.632984] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.633322] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] raise self.value [ 769.633322] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.633322] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] updated_port = self._update_port( [ 769.633322] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.633322] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] _ensure_no_port_binding_failure(port) [ 769.633322] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.633322] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] raise exception.PortBindingFailed(port_id=port['id']) [ 769.633322] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] nova.exception.PortBindingFailed: Binding failed for port dbed6212-8b1d-4e39-bcd7-b0cf04b60660, please check neutron logs for more information. [ 769.633322] env[63293]: ERROR nova.compute.manager [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] [ 769.633322] env[63293]: DEBUG nova.compute.utils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Binding failed for port dbed6212-8b1d-4e39-bcd7-b0cf04b60660, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 769.634609] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.755s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.637589] env[63293]: DEBUG nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Build of instance 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9 was re-scheduled: Binding failed for port dbed6212-8b1d-4e39-bcd7-b0cf04b60660, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 769.637960] env[63293]: DEBUG nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 769.638198] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Acquiring lock "refresh_cache-1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.638342] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Acquired lock "refresh_cache-1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.638522] env[63293]: DEBUG nova.network.neutron [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.854069] env[63293]: INFO nova.compute.manager [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fd67940d-9552-42b3-b472-2ce0e7899ea6] Took 1.02 seconds to deallocate network for instance. [ 769.966067] env[63293]: DEBUG nova.compute.manager [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 770.072943] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Releasing lock "refresh_cache-b5334832-25dc-4b03-9689-388a1382200a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.073443] env[63293]: DEBUG nova.compute.manager [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 770.073637] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 770.074645] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7424de-642c-4e4c-a1f9-fc08dc3522ea {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.083535] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 770.083535] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1554654e-721c-4da9-81d1-d3726f3cca47 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.089483] env[63293]: DEBUG oslo_vmware.api [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 770.089483] env[63293]: value = "task-1327502" [ 770.089483] env[63293]: _type = "Task" [ 770.089483] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.100423] env[63293]: DEBUG oslo_vmware.api [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327502, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.160831] env[63293]: DEBUG nova.network.neutron [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.250339] env[63293]: DEBUG nova.network.neutron [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.482954] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70057ba-dd05-484c-936c-9e8003414a97 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.490572] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.492510] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159c73ca-8ced-4e62-baed-8d45f7aebe13 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.523238] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61fa2fa4-9dec-4446-a2d8-91284481687d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.532090] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b1f784-2544-4404-8cee-09f987593407 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.546918] env[63293]: DEBUG nova.compute.provider_tree [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.604300] env[63293]: DEBUG oslo_vmware.api [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327502, 'name': PowerOffVM_Task, 'duration_secs': 0.208363} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.604640] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 770.604859] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 770.605123] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b3ecd5e-8ad2-49d3-bdd2-cd6823116596 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.638591] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 770.638826] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 770.639011] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Deleting the datastore file [datastore2] b5334832-25dc-4b03-9689-388a1382200a {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 770.639565] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04a37ee9-8916-468e-8e29-e83d37346c0a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.645363] env[63293]: DEBUG oslo_vmware.api [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for the task: (returnval){ [ 770.645363] env[63293]: value = "task-1327504" [ 770.645363] env[63293]: _type = "Task" [ 770.645363] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.655131] env[63293]: DEBUG oslo_vmware.api [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327504, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.755291] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Releasing lock "refresh_cache-1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.755532] env[63293]: DEBUG nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 770.755714] env[63293]: DEBUG nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.755963] env[63293]: DEBUG nova.network.neutron [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.774361] env[63293]: DEBUG nova.network.neutron [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.891982] env[63293]: INFO nova.scheduler.client.report [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleted allocations for instance fd67940d-9552-42b3-b472-2ce0e7899ea6 [ 771.052446] env[63293]: DEBUG nova.scheduler.client.report [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.156682] env[63293]: DEBUG oslo_vmware.api [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Task: {'id': task-1327504, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102676} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.156887] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 771.157083] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 771.157262] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.157429] env[63293]: INFO nova.compute.manager [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Took 1.08 seconds to destroy the instance on the hypervisor. [ 771.157663] env[63293]: DEBUG oslo.service.loopingcall [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.157848] env[63293]: DEBUG nova.compute.manager [-] [instance: b5334832-25dc-4b03-9689-388a1382200a] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.157960] env[63293]: DEBUG nova.network.neutron [-] [instance: b5334832-25dc-4b03-9689-388a1382200a] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 771.183316] env[63293]: DEBUG nova.network.neutron [-] [instance: b5334832-25dc-4b03-9689-388a1382200a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.276803] env[63293]: DEBUG nova.network.neutron [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.402776] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f6fcfca-803c-497e-8212-96b048c7b376 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "fd67940d-9552-42b3-b472-2ce0e7899ea6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.123s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.560944] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.926s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.561596] env[63293]: ERROR nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3b1821c8-1889-473b-8f69-e57fa4768f18, please check neutron logs for more information. [ 771.561596] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Traceback (most recent call last): [ 771.561596] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 771.561596] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] self.driver.spawn(context, instance, image_meta, [ 771.561596] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 771.561596] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] self._vmops.spawn(context, instance, image_meta, injected_files, [ 771.561596] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 771.561596] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] vm_ref = self.build_virtual_machine(instance, [ 771.561596] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 771.561596] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] vif_infos = vmwarevif.get_vif_info(self._session, [ 771.561596] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] for vif in network_info: [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] return self._sync_wrapper(fn, *args, **kwargs) [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] self.wait() [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] self[:] = self._gt.wait() [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] return self._exit_event.wait() [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] current.throw(*self._exc) [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 771.561924] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] result = function(*args, **kwargs) [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] return func(*args, **kwargs) [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] raise e [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] nwinfo = self.network_api.allocate_for_instance( [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] created_port_ids = self._update_ports_for_instance( [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] with excutils.save_and_reraise_exception(): [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] self.force_reraise() [ 771.562260] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.563216] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] raise self.value [ 771.563216] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.563216] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] updated_port = self._update_port( [ 771.563216] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.563216] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] _ensure_no_port_binding_failure(port) [ 771.563216] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.563216] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] raise exception.PortBindingFailed(port_id=port['id']) [ 771.563216] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] nova.exception.PortBindingFailed: Binding failed for port 3b1821c8-1889-473b-8f69-e57fa4768f18, please check neutron logs for more information. [ 771.563216] env[63293]: ERROR nova.compute.manager [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] [ 771.563216] env[63293]: DEBUG nova.compute.utils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Binding failed for port 3b1821c8-1889-473b-8f69-e57fa4768f18, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 771.563602] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.156s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.566457] env[63293]: DEBUG nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Build of instance 985a5c30-551f-4de7-bc85-d53f002e0f72 was re-scheduled: Binding failed for port 3b1821c8-1889-473b-8f69-e57fa4768f18, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 771.566903] env[63293]: DEBUG nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 771.567144] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Acquiring lock "refresh_cache-985a5c30-551f-4de7-bc85-d53f002e0f72" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.567291] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Acquired lock "refresh_cache-985a5c30-551f-4de7-bc85-d53f002e0f72" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.567444] env[63293]: DEBUG nova.network.neutron [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.687707] env[63293]: DEBUG nova.network.neutron [-] [instance: b5334832-25dc-4b03-9689-388a1382200a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.779627] env[63293]: INFO nova.compute.manager [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] [instance: 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9] Took 1.02 seconds to deallocate network for instance. [ 771.908593] env[63293]: DEBUG nova.compute.manager [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 772.093386] env[63293]: DEBUG nova.network.neutron [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.192185] env[63293]: INFO nova.compute.manager [-] [instance: b5334832-25dc-4b03-9689-388a1382200a] Took 1.03 seconds to deallocate network for instance. [ 772.224053] env[63293]: DEBUG nova.network.neutron [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.425128] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.469799] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d2cc5a-5fa1-497c-b9f3-096b326169ef {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.477519] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d12541e-11d6-458a-a602-23858b87e59c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.507200] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49bcef0-89a4-4b76-adfc-170e52545565 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.515419] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a049f78-3d06-486b-8c2e-981908f522cd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.528427] env[63293]: DEBUG nova.compute.provider_tree [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.700455] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.729653] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Releasing lock "refresh_cache-985a5c30-551f-4de7-bc85-d53f002e0f72" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.729653] env[63293]: DEBUG nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 772.729839] env[63293]: DEBUG nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.730032] env[63293]: DEBUG nova.network.neutron [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 772.748050] env[63293]: DEBUG nova.network.neutron [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.820082] env[63293]: INFO nova.scheduler.client.report [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Deleted allocations for instance 1d9c3eee-eefe-4a9a-b3d9-aac847892cc9 [ 773.032173] env[63293]: DEBUG nova.scheduler.client.report [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.251252] env[63293]: DEBUG nova.network.neutron [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.328161] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f66f328-bd55-45ea-9f08-0265e6f4deab tempest-ServerPasswordTestJSON-1852835389 tempest-ServerPasswordTestJSON-1852835389-project-member] Lock "1d9c3eee-eefe-4a9a-b3d9-aac847892cc9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.257s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.538340] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.538984] env[63293]: ERROR nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 536be720-3b98-49a5-a4c9-cfa83b018886, please check neutron logs for more information. [ 773.538984] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Traceback (most recent call last): [ 773.538984] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 773.538984] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] self.driver.spawn(context, instance, image_meta, [ 773.538984] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 773.538984] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.538984] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.538984] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] vm_ref = self.build_virtual_machine(instance, [ 773.538984] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.538984] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.538984] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] for vif in network_info: [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] return self._sync_wrapper(fn, *args, **kwargs) [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] self.wait() [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] self[:] = self._gt.wait() [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] return self._exit_event.wait() [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] current.throw(*self._exc) [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.539432] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] result = function(*args, **kwargs) [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] return func(*args, **kwargs) [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] raise e [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] nwinfo = self.network_api.allocate_for_instance( [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] created_port_ids = self._update_ports_for_instance( [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] with excutils.save_and_reraise_exception(): [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] self.force_reraise() [ 773.539966] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.540427] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] raise self.value [ 773.540427] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.540427] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] updated_port = self._update_port( [ 773.540427] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.540427] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] _ensure_no_port_binding_failure(port) [ 773.540427] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.540427] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] raise exception.PortBindingFailed(port_id=port['id']) [ 773.540427] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] nova.exception.PortBindingFailed: Binding failed for port 536be720-3b98-49a5-a4c9-cfa83b018886, please check neutron logs for more information. [ 773.540427] env[63293]: ERROR nova.compute.manager [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] [ 773.540427] env[63293]: DEBUG nova.compute.utils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Binding failed for port 536be720-3b98-49a5-a4c9-cfa83b018886, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 773.540939] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.366s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.544873] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Build of instance 67839a84-c1a0-4728-88f4-47acf99f1fc4 was re-scheduled: Binding failed for port 536be720-3b98-49a5-a4c9-cfa83b018886, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 773.545367] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 773.545646] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "refresh_cache-67839a84-c1a0-4728-88f4-47acf99f1fc4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.545856] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired lock "refresh_cache-67839a84-c1a0-4728-88f4-47acf99f1fc4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.546095] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.757471] env[63293]: INFO nova.compute.manager [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] [instance: 985a5c30-551f-4de7-bc85-d53f002e0f72] Took 1.02 seconds to deallocate network for instance. [ 773.811918] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.812381] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.833601] env[63293]: DEBUG nova.compute.manager [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 774.074445] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.190191] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.352260] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.425517] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80317e64-6fe1-4d2a-a62c-826d0d045ebe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.435148] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b911d5f3-fd2e-4164-b0e4-20fab3f2efad {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.474513] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1973d87d-abb2-49af-9655-5b84c7e37548 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.482097] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9d145c-18c3-4881-8594-a2f8e3a8cbf9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.495855] env[63293]: DEBUG nova.compute.provider_tree [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.693020] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Releasing lock "refresh_cache-67839a84-c1a0-4728-88f4-47acf99f1fc4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.693235] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 774.694040] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.694040] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.707703] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.789736] env[63293]: INFO nova.scheduler.client.report [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Deleted allocations for instance 985a5c30-551f-4de7-bc85-d53f002e0f72 [ 774.999209] env[63293]: DEBUG nova.scheduler.client.report [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.210727] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.296153] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b136c0a-3314-46c4-bca1-8d5573c7cd51 tempest-ServerTagsTestJSON-1821821476 tempest-ServerTagsTestJSON-1821821476-project-member] Lock "985a5c30-551f-4de7-bc85-d53f002e0f72" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.274s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.505967] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.965s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.506658] env[63293]: ERROR nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 861f3155-fa0f-452c-8005-af67c9fae028, please check neutron logs for more information. [ 775.506658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Traceback (most recent call last): [ 775.506658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.506658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] self.driver.spawn(context, instance, image_meta, [ 775.506658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 775.506658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.506658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.506658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] vm_ref = self.build_virtual_machine(instance, [ 775.506658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.506658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.506658] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] for vif in network_info: [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] return self._sync_wrapper(fn, *args, **kwargs) [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] self.wait() [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] self[:] = self._gt.wait() [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] return self._exit_event.wait() [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] current.throw(*self._exc) [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.507089] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] result = function(*args, **kwargs) [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] return func(*args, **kwargs) [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] raise e [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] nwinfo = self.network_api.allocate_for_instance( [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] created_port_ids = self._update_ports_for_instance( [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] with excutils.save_and_reraise_exception(): [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] self.force_reraise() [ 775.507869] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.508305] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] raise self.value [ 775.508305] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.508305] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] updated_port = self._update_port( [ 775.508305] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.508305] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] _ensure_no_port_binding_failure(port) [ 775.508305] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.508305] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] raise exception.PortBindingFailed(port_id=port['id']) [ 775.508305] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] nova.exception.PortBindingFailed: Binding failed for port 861f3155-fa0f-452c-8005-af67c9fae028, please check neutron logs for more information. [ 775.508305] env[63293]: ERROR nova.compute.manager [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] [ 775.508305] env[63293]: DEBUG nova.compute.utils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Binding failed for port 861f3155-fa0f-452c-8005-af67c9fae028, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 775.508852] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.072s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.510399] env[63293]: INFO nova.compute.claims [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.513070] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Build of instance 0df77939-d60c-4ccb-98f7-2325062ec76f was re-scheduled: Binding failed for port 861f3155-fa0f-452c-8005-af67c9fae028, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 775.516566] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 775.516566] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "refresh_cache-0df77939-d60c-4ccb-98f7-2325062ec76f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.516566] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired lock "refresh_cache-0df77939-d60c-4ccb-98f7-2325062ec76f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.516566] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.713843] env[63293]: INFO nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 67839a84-c1a0-4728-88f4-47acf99f1fc4] Took 1.02 seconds to deallocate network for instance. [ 775.800273] env[63293]: DEBUG nova.compute.manager [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 776.048993] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.152249] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.397270] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.544199] env[63293]: DEBUG nova.scheduler.client.report [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Refreshing inventories for resource provider 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 776.566823] env[63293]: DEBUG nova.scheduler.client.report [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Updating ProviderTree inventory for provider 619725c3-6194-4724-94b1-1dd499be55d0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 776.567100] env[63293]: DEBUG nova.compute.provider_tree [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 776.578411] env[63293]: DEBUG nova.scheduler.client.report [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Refreshing aggregate associations for resource provider 619725c3-6194-4724-94b1-1dd499be55d0, aggregates: None {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 776.597676] env[63293]: DEBUG nova.scheduler.client.report [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Refreshing trait associations for resource provider 619725c3-6194-4724-94b1-1dd499be55d0, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 776.655766] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Releasing lock "refresh_cache-0df77939-d60c-4ccb-98f7-2325062ec76f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.655974] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 776.656177] env[63293]: DEBUG nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.656344] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 776.675433] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.746195] env[63293]: INFO nova.scheduler.client.report [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Deleted allocations for instance 67839a84-c1a0-4728-88f4-47acf99f1fc4 [ 776.955162] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f416f1-24af-4972-8c30-2097f2783fde {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.963011] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27f6784-ecb2-442b-99b6-476d500c735f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.994029] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-901725b7-6cbe-4555-8196-5c2c2c678b1e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.001953] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77f732f-2cf9-41d3-9da0-4418c0c9a97f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.017945] env[63293]: DEBUG nova.compute.provider_tree [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.180213] env[63293]: DEBUG nova.network.neutron [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.257874] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "67839a84-c1a0-4728-88f4-47acf99f1fc4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.184s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.520992] env[63293]: DEBUG nova.scheduler.client.report [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.685327] env[63293]: INFO nova.compute.manager [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 0df77939-d60c-4ccb-98f7-2325062ec76f] Took 1.03 seconds to deallocate network for instance. [ 777.762459] env[63293]: DEBUG nova.compute.manager [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 778.026113] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.026691] env[63293]: DEBUG nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 778.029278] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.106s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.284616] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.537540] env[63293]: DEBUG nova.compute.utils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 778.539153] env[63293]: DEBUG nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 778.540163] env[63293]: DEBUG nova.network.neutron [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 778.604347] env[63293]: DEBUG nova.policy [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0aa775dd2d2a4777a92f6b99668ee89f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad88e731209c49d3bb24e024c3a72865', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 778.715749] env[63293]: INFO nova.scheduler.client.report [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Deleted allocations for instance 0df77939-d60c-4ccb-98f7-2325062ec76f [ 778.843393] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0664b8e7-d934-44d8-a4af-9ea99847ff3a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.851420] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4799fd43-3b2b-40bc-997a-f8791a88e178 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.883079] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cdeecbf-d95b-46d2-9de4-21f87765bc48 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.890442] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4caad18-1824-4ffc-b266-c4e46bdb3e7b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.903686] env[63293]: DEBUG nova.compute.provider_tree [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.929299] env[63293]: DEBUG nova.network.neutron [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Successfully created port: 9276c7bd-1873-447d-aa82-96375b594c8b {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 779.046110] env[63293]: DEBUG nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 779.227935] env[63293]: DEBUG oslo_concurrency.lockutils [None req-931afe0b-4e45-4b9a-9539-948e1cd9f92a tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "0df77939-d60c-4ccb-98f7-2325062ec76f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.123s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.412227] env[63293]: DEBUG nova.scheduler.client.report [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.733599] env[63293]: DEBUG nova.compute.manager [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.820639] env[63293]: DEBUG nova.compute.manager [req-e8019ce4-3378-492b-a6de-310f03e5adba req-488fef40-54ab-48d3-92ba-9aac5ab4655a service nova] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Received event network-changed-9276c7bd-1873-447d-aa82-96375b594c8b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.820836] env[63293]: DEBUG nova.compute.manager [req-e8019ce4-3378-492b-a6de-310f03e5adba req-488fef40-54ab-48d3-92ba-9aac5ab4655a service nova] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Refreshing instance network info cache due to event network-changed-9276c7bd-1873-447d-aa82-96375b594c8b. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 779.821067] env[63293]: DEBUG oslo_concurrency.lockutils [req-e8019ce4-3378-492b-a6de-310f03e5adba req-488fef40-54ab-48d3-92ba-9aac5ab4655a service nova] Acquiring lock "refresh_cache-52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.821212] env[63293]: DEBUG oslo_concurrency.lockutils [req-e8019ce4-3378-492b-a6de-310f03e5adba req-488fef40-54ab-48d3-92ba-9aac5ab4655a service nova] Acquired lock "refresh_cache-52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.821369] env[63293]: DEBUG nova.network.neutron [req-e8019ce4-3378-492b-a6de-310f03e5adba req-488fef40-54ab-48d3-92ba-9aac5ab4655a service nova] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Refreshing network info cache for port 9276c7bd-1873-447d-aa82-96375b594c8b {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 779.916742] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.887s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.917424] env[63293]: ERROR nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f1003952-8806-4d1c-9b8d-007aa7ea5792, please check neutron logs for more information. [ 779.917424] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Traceback (most recent call last): [ 779.917424] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.917424] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] self.driver.spawn(context, instance, image_meta, [ 779.917424] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 779.917424] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.917424] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.917424] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] vm_ref = self.build_virtual_machine(instance, [ 779.917424] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.917424] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.917424] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] for vif in network_info: [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] return self._sync_wrapper(fn, *args, **kwargs) [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] self.wait() [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] self[:] = self._gt.wait() [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] return self._exit_event.wait() [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] result = hub.switch() [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 779.917811] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] return self.greenlet.switch() [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] result = function(*args, **kwargs) [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] return func(*args, **kwargs) [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] raise e [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] nwinfo = self.network_api.allocate_for_instance( [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] created_port_ids = self._update_ports_for_instance( [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] with excutils.save_and_reraise_exception(): [ 779.918107] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.918380] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] self.force_reraise() [ 779.918380] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.918380] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] raise self.value [ 779.918380] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.918380] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] updated_port = self._update_port( [ 779.918380] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.918380] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] _ensure_no_port_binding_failure(port) [ 779.918380] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.918380] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] raise exception.PortBindingFailed(port_id=port['id']) [ 779.918380] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] nova.exception.PortBindingFailed: Binding failed for port f1003952-8806-4d1c-9b8d-007aa7ea5792, please check neutron logs for more information. [ 779.918380] env[63293]: ERROR nova.compute.manager [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] [ 779.918667] env[63293]: DEBUG nova.compute.utils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Binding failed for port f1003952-8806-4d1c-9b8d-007aa7ea5792, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 779.920616] env[63293]: DEBUG nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Build of instance b1b1ebe0-187b-4e23-9433-7dff420e3de3 was re-scheduled: Binding failed for port f1003952-8806-4d1c-9b8d-007aa7ea5792, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 779.921078] env[63293]: DEBUG nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 779.921309] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Acquiring lock "refresh_cache-b1b1ebe0-187b-4e23-9433-7dff420e3de3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.921455] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Acquired lock "refresh_cache-b1b1ebe0-187b-4e23-9433-7dff420e3de3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.921614] env[63293]: DEBUG nova.network.neutron [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.922636] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.055s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.927237] env[63293]: ERROR nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9276c7bd-1873-447d-aa82-96375b594c8b, please check neutron logs for more information. [ 779.927237] env[63293]: ERROR nova.compute.manager Traceback (most recent call last): [ 779.927237] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.927237] env[63293]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 779.927237] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.927237] env[63293]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 779.927237] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.927237] env[63293]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 779.927237] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.927237] env[63293]: ERROR nova.compute.manager self.force_reraise() [ 779.927237] env[63293]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.927237] env[63293]: ERROR nova.compute.manager raise self.value [ 779.927237] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.927237] env[63293]: ERROR nova.compute.manager updated_port = self._update_port( [ 779.927237] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.927237] env[63293]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 779.927650] env[63293]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.927650] env[63293]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 779.927650] env[63293]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9276c7bd-1873-447d-aa82-96375b594c8b, please check neutron logs for more information. [ 779.927650] env[63293]: ERROR nova.compute.manager [ 779.927650] env[63293]: Traceback (most recent call last): [ 779.927650] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 779.927650] env[63293]: listener.cb(fileno) [ 779.927650] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.927650] env[63293]: result = function(*args, **kwargs) [ 779.927650] env[63293]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.927650] env[63293]: return func(*args, **kwargs) [ 779.927650] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.927650] env[63293]: raise e [ 779.927650] env[63293]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.927650] env[63293]: nwinfo = self.network_api.allocate_for_instance( [ 779.927650] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.927650] env[63293]: created_port_ids = self._update_ports_for_instance( [ 779.927650] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.927650] env[63293]: with excutils.save_and_reraise_exception(): [ 779.927650] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.927650] env[63293]: self.force_reraise() [ 779.927650] env[63293]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.927650] env[63293]: raise self.value [ 779.927650] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.927650] env[63293]: updated_port = self._update_port( [ 779.927650] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.927650] env[63293]: _ensure_no_port_binding_failure(port) [ 779.927650] env[63293]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.927650] env[63293]: raise exception.PortBindingFailed(port_id=port['id']) [ 779.928323] env[63293]: nova.exception.PortBindingFailed: Binding failed for port 9276c7bd-1873-447d-aa82-96375b594c8b, please check neutron logs for more information. [ 779.928323] env[63293]: Removing descriptor: 15 [ 780.060663] env[63293]: DEBUG nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 780.090518] env[63293]: DEBUG nova.virt.hardware [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 780.090801] env[63293]: DEBUG nova.virt.hardware [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 780.090962] env[63293]: DEBUG nova.virt.hardware [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.091156] env[63293]: DEBUG nova.virt.hardware [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 780.091300] env[63293]: DEBUG nova.virt.hardware [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.091484] env[63293]: DEBUG nova.virt.hardware [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 780.091667] env[63293]: DEBUG nova.virt.hardware [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 780.091794] env[63293]: DEBUG nova.virt.hardware [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 780.092615] env[63293]: DEBUG nova.virt.hardware [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 780.092615] env[63293]: DEBUG nova.virt.hardware [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 780.092615] env[63293]: DEBUG nova.virt.hardware [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.093157] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f2fe6c-2feb-4c7e-8165-8d6fe83d60e5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.101276] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d9c040-f283-475c-bf0d-f00505bb0bc7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.117033] env[63293]: ERROR nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9276c7bd-1873-447d-aa82-96375b594c8b, please check neutron logs for more information. [ 780.117033] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Traceback (most recent call last): [ 780.117033] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 780.117033] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] yield resources [ 780.117033] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 780.117033] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] self.driver.spawn(context, instance, image_meta, [ 780.117033] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 780.117033] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 780.117033] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 780.117033] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] vm_ref = self.build_virtual_machine(instance, [ 780.117033] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] vif_infos = vmwarevif.get_vif_info(self._session, [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] for vif in network_info: [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] return self._sync_wrapper(fn, *args, **kwargs) [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] self.wait() [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] self[:] = self._gt.wait() [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] return self._exit_event.wait() [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 780.117667] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] current.throw(*self._exc) [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] result = function(*args, **kwargs) [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] return func(*args, **kwargs) [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] raise e [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] nwinfo = self.network_api.allocate_for_instance( [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] created_port_ids = self._update_ports_for_instance( [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] with excutils.save_and_reraise_exception(): [ 780.118067] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 780.118396] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] self.force_reraise() [ 780.118396] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 780.118396] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] raise self.value [ 780.118396] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 780.118396] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] updated_port = self._update_port( [ 780.118396] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 780.118396] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] _ensure_no_port_binding_failure(port) [ 780.118396] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 780.118396] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] raise exception.PortBindingFailed(port_id=port['id']) [ 780.118396] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] nova.exception.PortBindingFailed: Binding failed for port 9276c7bd-1873-447d-aa82-96375b594c8b, please check neutron logs for more information. [ 780.118396] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] [ 780.118396] env[63293]: INFO nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Terminating instance [ 780.118945] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Acquiring lock "refresh_cache-52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.256070] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.338148] env[63293]: DEBUG nova.network.neutron [req-e8019ce4-3378-492b-a6de-310f03e5adba req-488fef40-54ab-48d3-92ba-9aac5ab4655a service nova] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.437429] env[63293]: DEBUG nova.network.neutron [req-e8019ce4-3378-492b-a6de-310f03e5adba req-488fef40-54ab-48d3-92ba-9aac5ab4655a service nova] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.449714] env[63293]: DEBUG nova.network.neutron [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.543134] env[63293]: DEBUG nova.network.neutron [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.943039] env[63293]: DEBUG oslo_concurrency.lockutils [req-e8019ce4-3378-492b-a6de-310f03e5adba req-488fef40-54ab-48d3-92ba-9aac5ab4655a service nova] Releasing lock "refresh_cache-52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.943039] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Acquired lock "refresh_cache-52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.943039] env[63293]: DEBUG nova.network.neutron [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.957085] env[63293]: WARNING nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance b5334832-25dc-4b03-9689-388a1382200a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 781.045718] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Releasing lock "refresh_cache-b1b1ebe0-187b-4e23-9433-7dff420e3de3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.046040] env[63293]: DEBUG nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 781.046224] env[63293]: DEBUG nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 781.046488] env[63293]: DEBUG nova.network.neutron [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 781.061160] env[63293]: DEBUG nova.network.neutron [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.101825] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.102101] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.129248] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.129248] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.460159] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance b1b1ebe0-187b-4e23-9433-7dff420e3de3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 781.460335] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 781.460461] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 781.465311] env[63293]: DEBUG nova.network.neutron [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.561941] env[63293]: DEBUG nova.network.neutron [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.563301] env[63293]: DEBUG nova.network.neutron [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.847782] env[63293]: DEBUG nova.compute.manager [req-a50a49bc-1247-45fd-9023-bebfa8586af8 req-1a996176-2f50-4481-9b1d-9ab4f99b89f4 service nova] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Received event network-vif-deleted-9276c7bd-1873-447d-aa82-96375b594c8b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.963238] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance b48324a4-902d-4642-8827-daee9683b3a0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 782.065501] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Releasing lock "refresh_cache-52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.065925] env[63293]: DEBUG nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 782.066135] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 782.066435] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab801d74-260b-4a2b-a958-b9a6c6467516 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.068730] env[63293]: INFO nova.compute.manager [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] [instance: b1b1ebe0-187b-4e23-9433-7dff420e3de3] Took 1.02 seconds to deallocate network for instance. [ 782.078057] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c305aeed-a496-4ab9-867b-ea36fb330808 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.100996] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4 could not be found. [ 782.101242] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.101698] env[63293]: INFO nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 782.101698] env[63293]: DEBUG oslo.service.loopingcall [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.101860] env[63293]: DEBUG nova.compute.manager [-] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.101957] env[63293]: DEBUG nova.network.neutron [-] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.117317] env[63293]: DEBUG nova.network.neutron [-] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.926845] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 495cb1b0-9194-44e6-9f68-de62033ea17d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 782.930796] env[63293]: DEBUG nova.network.neutron [-] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.433068] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 783.436948] env[63293]: INFO nova.compute.manager [-] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Took 1.33 seconds to deallocate network for instance. [ 783.439452] env[63293]: DEBUG nova.compute.claims [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Aborting claim: {{(pid=63293) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 783.439641] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.459279] env[63293]: INFO nova.scheduler.client.report [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Deleted allocations for instance b1b1ebe0-187b-4e23-9433-7dff420e3de3 [ 783.938991] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 783.968158] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a6361c1f-c092-48e4-947a-8ee9edfe5dc4 tempest-InstanceActionsV221TestJSON-33448373 tempest-InstanceActionsV221TestJSON-33448373-project-member] Lock "b1b1ebe0-187b-4e23-9433-7dff420e3de3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.126s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.442017] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance c8009fc0-f7dc-4568-9909-7fa9f7f6a56c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 784.471945] env[63293]: DEBUG nova.compute.manager [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.944986] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 65c2f72d-6bbe-45ac-8efc-401e0664390f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 784.998545] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.448635] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 33d71260-26f1-482c-b93b-5f8e27c545f3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 785.951864] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance b8637881-9fb6-4c5d-848d-7b2d38f8e970 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 786.456161] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 48603812-c3cc-4dae-b8ba-99b9ac5f7969 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 786.959810] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance ad585ebb-2072-45df-b645-94c9fa93576b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.466555] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 831d7d9b-364f-4c29-bda5-e6c38291a973 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.970049] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 1e01a75d-122d-4122-9fed-4164d64d4ee8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.474043] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 4593cabd-1331-4ee1-99f5-46edeac56859 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.976738] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 93080446-fd45-4281-bf6e-35d1d964f41e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.479605] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance c26d6a3a-3643-4b52-870a-4e2ed901d908 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.982397] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 9ddf4f23-1279-4cbd-8212-10f344060445 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.485701] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance f427933b-dd2c-49a9-b401-a647635a1eab has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.989275] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 6cd8b0eb-4bfe-49e0-b392-900f50865ad9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.496233] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 8ea7f3c3-8d73-4fb2-a848-4be4649e5660 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.000067] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 9973dd7f-ca75-4db0-8d2b-f21b0311abeb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.000067] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 792.000067] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 792.262016] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54b5c6f-7607-4bf0-bf22-49900e7ac2e6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.268437] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82dddd40-db1c-41ba-ae28-299f47e8acbe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.304233] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a0bcf0-e1b5-4b64-89ff-0361bb6c6230 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.314740] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c00d7b3-cce1-44e7-a412-94560f4d8a41 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.325574] env[63293]: DEBUG nova.compute.provider_tree [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.828422] env[63293]: DEBUG nova.scheduler.client.report [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.332835] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 793.333148] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.410s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.333432] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.843s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.336415] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.336622] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Cleaning up deleted instances {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 793.844357] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] There are 3 instances to clean {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 793.844626] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 0aade4da-882a-4342-aefb-07eb306b17ff] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 794.083281] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa991b2-4d26-44a4-97ae-bd3ae93d6fef {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.090703] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24464c8d-e08f-40bc-b22f-25f6ecc0ac06 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.119566] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9177f505-a0a0-47c5-95ed-18ac3e94a9b7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.126167] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666a9d26-9dcd-43f4-bb8e-a9ca95e18279 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.138986] env[63293]: DEBUG nova.compute.provider_tree [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.347698] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: b2055187-edd8-4a8c-9811-183cf7ba3b39] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 794.641923] env[63293]: DEBUG nova.scheduler.client.report [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.850441] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: e80209b3-273c-4045-ad8a-2a3c4c48200a] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 795.147293] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.814s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.147947] env[63293]: ERROR nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7707013f-9522-41aa-b0a5-c3b6d07a8bf3, please check neutron logs for more information. [ 795.147947] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Traceback (most recent call last): [ 795.147947] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 795.147947] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] self.driver.spawn(context, instance, image_meta, [ 795.147947] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 795.147947] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 795.147947] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 795.147947] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] vm_ref = self.build_virtual_machine(instance, [ 795.147947] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 795.147947] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] vif_infos = vmwarevif.get_vif_info(self._session, [ 795.147947] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] for vif in network_info: [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] return self._sync_wrapper(fn, *args, **kwargs) [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] self.wait() [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] self[:] = self._gt.wait() [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] return self._exit_event.wait() [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] current.throw(*self._exc) [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 795.148445] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] result = function(*args, **kwargs) [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] return func(*args, **kwargs) [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] raise e [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] nwinfo = self.network_api.allocate_for_instance( [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] created_port_ids = self._update_ports_for_instance( [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] with excutils.save_and_reraise_exception(): [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] self.force_reraise() [ 795.148790] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.149110] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] raise self.value [ 795.149110] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.149110] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] updated_port = self._update_port( [ 795.149110] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.149110] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] _ensure_no_port_binding_failure(port) [ 795.149110] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.149110] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] raise exception.PortBindingFailed(port_id=port['id']) [ 795.149110] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] nova.exception.PortBindingFailed: Binding failed for port 7707013f-9522-41aa-b0a5-c3b6d07a8bf3, please check neutron logs for more information. [ 795.149110] env[63293]: ERROR nova.compute.manager [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] [ 795.149110] env[63293]: DEBUG nova.compute.utils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Binding failed for port 7707013f-9522-41aa-b0a5-c3b6d07a8bf3, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 795.150745] env[63293]: DEBUG nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Build of instance 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a was re-scheduled: Binding failed for port 7707013f-9522-41aa-b0a5-c3b6d07a8bf3, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 795.151227] env[63293]: DEBUG nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 795.151456] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Acquiring lock "refresh_cache-8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.151602] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Acquired lock "refresh_cache-8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.151758] env[63293]: DEBUG nova.network.neutron [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.152767] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.541s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.154563] env[63293]: INFO nova.compute.claims [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.353829] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 795.354061] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Cleaning up deleted instances with incomplete migration {{(pid=63293) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 795.677361] env[63293]: DEBUG nova.network.neutron [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.770251] env[63293]: DEBUG nova.network.neutron [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.857742] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 796.273061] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Releasing lock "refresh_cache-8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.273290] env[63293]: DEBUG nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 796.273471] env[63293]: DEBUG nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.273637] env[63293]: DEBUG nova.network.neutron [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.290706] env[63293]: DEBUG nova.network.neutron [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.443907] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f34b18-9cf4-40fd-82b0-747695883870 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.451483] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159d0172-658f-43d7-993d-0af75397e3f6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.482039] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc38e70-1148-41c7-a952-2e612949e539 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.488998] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266427c8-ebd4-4152-9a0f-e52fbcd87471 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.501808] env[63293]: DEBUG nova.compute.provider_tree [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.794092] env[63293]: DEBUG nova.network.neutron [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.006018] env[63293]: DEBUG nova.scheduler.client.report [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.299823] env[63293]: INFO nova.compute.manager [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] [instance: 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a] Took 1.03 seconds to deallocate network for instance. [ 797.510107] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.510662] env[63293]: DEBUG nova.compute.manager [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 797.513180] env[63293]: DEBUG oslo_concurrency.lockutils [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.604s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.513355] env[63293]: DEBUG nova.objects.instance [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] [instance: b5334832-25dc-4b03-9689-388a1382200a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63293) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 798.017529] env[63293]: DEBUG nova.compute.utils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 798.019511] env[63293]: DEBUG nova.compute.manager [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 798.019802] env[63293]: DEBUG nova.network.neutron [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 798.062429] env[63293]: DEBUG nova.policy [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b71c6c0cc133419ea1df4ff2ebba1d8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b67ae2932d741c4b3a03d5d344127ef', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 798.326805] env[63293]: INFO nova.scheduler.client.report [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Deleted allocations for instance 8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a [ 798.349104] env[63293]: DEBUG nova.network.neutron [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Successfully created port: c3bd127f-e525-4481-9814-75afe260e314 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.520512] env[63293]: DEBUG nova.compute.manager [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 798.526032] env[63293]: DEBUG oslo_concurrency.lockutils [None req-175dcc0e-898e-47f7-bad1-801cc44f023d tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.526972] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.036s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.530064] env[63293]: INFO nova.compute.claims [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.835384] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a727d98-474f-4822-a9cc-68f354d3fb2d tempest-ServersTestBootFromVolume-2134488671 tempest-ServersTestBootFromVolume-2134488671-project-member] Lock "8ddbb8f2-e8f0-44ed-a095-c1f48d4b174a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.746s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.339582] env[63293]: DEBUG nova.compute.manager [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 799.530637] env[63293]: DEBUG nova.compute.manager [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 799.562068] env[63293]: DEBUG nova.virt.hardware [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 799.562348] env[63293]: DEBUG nova.virt.hardware [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 799.562552] env[63293]: DEBUG nova.virt.hardware [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.562768] env[63293]: DEBUG nova.virt.hardware [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 799.562946] env[63293]: DEBUG nova.virt.hardware [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.563136] env[63293]: DEBUG nova.virt.hardware [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 799.563349] env[63293]: DEBUG nova.virt.hardware [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 799.563511] env[63293]: DEBUG nova.virt.hardware [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 799.563684] env[63293]: DEBUG nova.virt.hardware [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 799.563871] env[63293]: DEBUG nova.virt.hardware [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 799.564070] env[63293]: DEBUG nova.virt.hardware [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.564921] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c00f7b9-4ff6-4a7d-8d17-e161f1ea7b18 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.576363] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07e5663-f680-4104-9944-9bae9e658129 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.824179] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee42970c-7cd9-4cbe-8373-b5b97143d775 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.833302] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1883bd7d-0eea-4e39-86b5-154fca1e44d6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.867057] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-706a4b93-d66b-494a-9d68-b84c1d1a73a4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.880756] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22a1619-8196-4907-b287-d8360997fb05 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.895683] env[63293]: DEBUG nova.compute.provider_tree [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.897772] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.907182] env[63293]: DEBUG nova.compute.manager [req-ad5f037d-b337-43e5-a3f5-655e02d3fc9e req-a9252933-1242-4a71-8c35-a2016c207e7b service nova] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Received event network-vif-plugged-c3bd127f-e525-4481-9814-75afe260e314 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.907182] env[63293]: DEBUG oslo_concurrency.lockutils [req-ad5f037d-b337-43e5-a3f5-655e02d3fc9e req-a9252933-1242-4a71-8c35-a2016c207e7b service nova] Acquiring lock "b48324a4-902d-4642-8827-daee9683b3a0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.907182] env[63293]: DEBUG oslo_concurrency.lockutils [req-ad5f037d-b337-43e5-a3f5-655e02d3fc9e req-a9252933-1242-4a71-8c35-a2016c207e7b service nova] Lock "b48324a4-902d-4642-8827-daee9683b3a0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.907182] env[63293]: DEBUG oslo_concurrency.lockutils [req-ad5f037d-b337-43e5-a3f5-655e02d3fc9e req-a9252933-1242-4a71-8c35-a2016c207e7b service nova] Lock "b48324a4-902d-4642-8827-daee9683b3a0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.907182] env[63293]: DEBUG nova.compute.manager [req-ad5f037d-b337-43e5-a3f5-655e02d3fc9e req-a9252933-1242-4a71-8c35-a2016c207e7b service nova] [instance: b48324a4-902d-4642-8827-daee9683b3a0] No waiting events found dispatching network-vif-plugged-c3bd127f-e525-4481-9814-75afe260e314 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.907335] env[63293]: WARNING nova.compute.manager [req-ad5f037d-b337-43e5-a3f5-655e02d3fc9e req-a9252933-1242-4a71-8c35-a2016c207e7b service nova] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Received unexpected event network-vif-plugged-c3bd127f-e525-4481-9814-75afe260e314 for instance with vm_state building and task_state spawning. [ 800.003154] env[63293]: DEBUG nova.network.neutron [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Successfully updated port: c3bd127f-e525-4481-9814-75afe260e314 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 800.403108] env[63293]: DEBUG nova.scheduler.client.report [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.507728] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "refresh_cache-b48324a4-902d-4642-8827-daee9683b3a0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.507728] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquired lock "refresh_cache-b48324a4-902d-4642-8827-daee9683b3a0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.507728] env[63293]: DEBUG nova.network.neutron [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 800.909245] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.909245] env[63293]: DEBUG nova.compute.manager [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.912621] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.488s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.915105] env[63293]: INFO nova.compute.claims [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.064251] env[63293]: DEBUG nova.network.neutron [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.300584] env[63293]: DEBUG nova.network.neutron [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Updating instance_info_cache with network_info: [{"id": "c3bd127f-e525-4481-9814-75afe260e314", "address": "fa:16:3e:b3:7e:19", "network": {"id": "8fa78b94-cb99-4538-82c2-24bd65b49e2e", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1527619107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b67ae2932d741c4b3a03d5d344127ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "664c466b-9417-49d7-83cc-364d964c403a", "external-id": "nsx-vlan-transportzone-103", "segmentation_id": 103, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3bd127f-e5", "ovs_interfaceid": "c3bd127f-e525-4481-9814-75afe260e314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.426978] env[63293]: DEBUG nova.compute.utils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.430529] env[63293]: DEBUG nova.compute.manager [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 801.430726] env[63293]: DEBUG nova.network.neutron [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 801.468941] env[63293]: DEBUG nova.policy [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af33165b77294edfbffc57caec100d3f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a3bcac63f7844e10a166eca717f910a1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.762176] env[63293]: DEBUG nova.network.neutron [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Successfully created port: 70383439-e38d-4f62-8f53-350c7b430342 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.801899] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Releasing lock "refresh_cache-b48324a4-902d-4642-8827-daee9683b3a0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.802318] env[63293]: DEBUG nova.compute.manager [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Instance network_info: |[{"id": "c3bd127f-e525-4481-9814-75afe260e314", "address": "fa:16:3e:b3:7e:19", "network": {"id": "8fa78b94-cb99-4538-82c2-24bd65b49e2e", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1527619107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b67ae2932d741c4b3a03d5d344127ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "664c466b-9417-49d7-83cc-364d964c403a", "external-id": "nsx-vlan-transportzone-103", "segmentation_id": 103, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3bd127f-e5", "ovs_interfaceid": "c3bd127f-e525-4481-9814-75afe260e314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 801.802817] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:7e:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '664c466b-9417-49d7-83cc-364d964c403a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c3bd127f-e525-4481-9814-75afe260e314', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.811126] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Creating folder: Project (3b67ae2932d741c4b3a03d5d344127ef). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.811303] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f13bfcda-8df3-4e8f-abe7-50ba51670910 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.823997] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Created folder: Project (3b67ae2932d741c4b3a03d5d344127ef) in parent group-v283678. [ 801.824215] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Creating folder: Instances. Parent ref: group-v283702. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.824454] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35edb891-9d57-4b8a-8ff1-b33373ea2249 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.832927] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Created folder: Instances in parent group-v283702. [ 801.833224] env[63293]: DEBUG oslo.service.loopingcall [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.833433] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 801.833636] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54576ea5-b247-4627-b950-aa53c02fc891 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.852368] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.852368] env[63293]: value = "task-1327508" [ 801.852368] env[63293]: _type = "Task" [ 801.852368] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.861269] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327508, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.931288] env[63293]: DEBUG nova.compute.manager [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 802.085494] env[63293]: DEBUG nova.compute.manager [req-d4f2f4e2-85ee-4960-ae6c-af1ffdc2744d req-370dff69-10e4-4697-9c6e-2b46c91e4f21 service nova] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Received event network-changed-c3bd127f-e525-4481-9814-75afe260e314 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.085723] env[63293]: DEBUG nova.compute.manager [req-d4f2f4e2-85ee-4960-ae6c-af1ffdc2744d req-370dff69-10e4-4697-9c6e-2b46c91e4f21 service nova] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Refreshing instance network info cache due to event network-changed-c3bd127f-e525-4481-9814-75afe260e314. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 802.086064] env[63293]: DEBUG oslo_concurrency.lockutils [req-d4f2f4e2-85ee-4960-ae6c-af1ffdc2744d req-370dff69-10e4-4697-9c6e-2b46c91e4f21 service nova] Acquiring lock "refresh_cache-b48324a4-902d-4642-8827-daee9683b3a0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.087072] env[63293]: DEBUG oslo_concurrency.lockutils [req-d4f2f4e2-85ee-4960-ae6c-af1ffdc2744d req-370dff69-10e4-4697-9c6e-2b46c91e4f21 service nova] Acquired lock "refresh_cache-b48324a4-902d-4642-8827-daee9683b3a0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.087072] env[63293]: DEBUG nova.network.neutron [req-d4f2f4e2-85ee-4960-ae6c-af1ffdc2744d req-370dff69-10e4-4697-9c6e-2b46c91e4f21 service nova] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Refreshing network info cache for port c3bd127f-e525-4481-9814-75afe260e314 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.290611] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb0e182-85f3-4b50-bdb7-1777688a65c3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.298119] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db1a364-0568-46f3-97ca-61581775a43c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.328241] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69b4885-b6d5-4e3a-b1d4-1359c81409e8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.335510] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca7b90c-d249-4127-b750-5d1eaa72d9de {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.348658] env[63293]: DEBUG nova.compute.provider_tree [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.360718] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327508, 'name': CreateVM_Task, 'duration_secs': 0.287465} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.361437] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 802.370211] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.370379] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.370690] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 802.371234] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dae887f1-8b1f-4037-88b6-2ff8a874717b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.375857] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for the task: (returnval){ [ 802.375857] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5283bee6-c1dd-f560-2e19-140e8677e7fb" [ 802.375857] env[63293]: _type = "Task" [ 802.375857] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.384386] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5283bee6-c1dd-f560-2e19-140e8677e7fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.851904] env[63293]: DEBUG nova.scheduler.client.report [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.888916] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5283bee6-c1dd-f560-2e19-140e8677e7fb, 'name': SearchDatastore_Task, 'duration_secs': 0.011169} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.891433] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.892027] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.892027] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.892027] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.892193] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.892417] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea4f6325-9971-4cd4-9838-d8aed108d2db {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.900322] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.900500] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 802.901169] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9ad318c-e2e2-47c1-9b8b-27248c1ec244 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.906165] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for the task: (returnval){ [ 802.906165] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52be0573-58e2-2101-a82b-d0d2619591b5" [ 802.906165] env[63293]: _type = "Task" [ 802.906165] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.913023] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52be0573-58e2-2101-a82b-d0d2619591b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.956064] env[63293]: DEBUG nova.compute.manager [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.981097] env[63293]: DEBUG nova.virt.hardware [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.981097] env[63293]: DEBUG nova.virt.hardware [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.981097] env[63293]: DEBUG nova.virt.hardware [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.981263] env[63293]: DEBUG nova.virt.hardware [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.981263] env[63293]: DEBUG nova.virt.hardware [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.981263] env[63293]: DEBUG nova.virt.hardware [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.981263] env[63293]: DEBUG nova.virt.hardware [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.981263] env[63293]: DEBUG nova.virt.hardware [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.981418] env[63293]: DEBUG nova.virt.hardware [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.982260] env[63293]: DEBUG nova.virt.hardware [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.982598] env[63293]: DEBUG nova.virt.hardware [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.983548] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83001c54-42c3-470f-a09e-41d0cf55365f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.992664] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a204de3-dca4-401f-8e9c-01520a5cd40d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.000021] env[63293]: DEBUG nova.network.neutron [req-d4f2f4e2-85ee-4960-ae6c-af1ffdc2744d req-370dff69-10e4-4697-9c6e-2b46c91e4f21 service nova] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Updated VIF entry in instance network info cache for port c3bd127f-e525-4481-9814-75afe260e314. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 803.000021] env[63293]: DEBUG nova.network.neutron [req-d4f2f4e2-85ee-4960-ae6c-af1ffdc2744d req-370dff69-10e4-4697-9c6e-2b46c91e4f21 service nova] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Updating instance_info_cache with network_info: [{"id": "c3bd127f-e525-4481-9814-75afe260e314", "address": "fa:16:3e:b3:7e:19", "network": {"id": "8fa78b94-cb99-4538-82c2-24bd65b49e2e", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1527619107-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b67ae2932d741c4b3a03d5d344127ef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "664c466b-9417-49d7-83cc-364d964c403a", "external-id": "nsx-vlan-transportzone-103", "segmentation_id": 103, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3bd127f-e5", "ovs_interfaceid": "c3bd127f-e525-4481-9814-75afe260e314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.360330] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.360330] env[63293]: DEBUG nova.compute.manager [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.360791] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.660s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.361152] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.364156] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.012s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.365993] env[63293]: INFO nova.compute.claims [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.402327] env[63293]: INFO nova.scheduler.client.report [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Deleted allocations for instance b5334832-25dc-4b03-9689-388a1382200a [ 803.420656] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52be0573-58e2-2101-a82b-d0d2619591b5, 'name': SearchDatastore_Task, 'duration_secs': 0.008504} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.421364] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-570b6274-fd28-445e-bec4-a03f717929b3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.428177] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for the task: (returnval){ [ 803.428177] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528fd2f9-d717-5204-c72f-1e2ce68a5a2b" [ 803.428177] env[63293]: _type = "Task" [ 803.428177] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.436341] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528fd2f9-d717-5204-c72f-1e2ce68a5a2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.514221] env[63293]: DEBUG oslo_concurrency.lockutils [req-d4f2f4e2-85ee-4960-ae6c-af1ffdc2744d req-370dff69-10e4-4697-9c6e-2b46c91e4f21 service nova] Releasing lock "refresh_cache-b48324a4-902d-4642-8827-daee9683b3a0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.631098] env[63293]: DEBUG nova.network.neutron [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Successfully updated port: 70383439-e38d-4f62-8f53-350c7b430342 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.872454] env[63293]: DEBUG nova.compute.utils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.877869] env[63293]: DEBUG nova.compute.manager [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.877869] env[63293]: DEBUG nova.network.neutron [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.919595] env[63293]: DEBUG oslo_concurrency.lockutils [None req-187f19ff-dba7-47fc-a125-df923a4b4b3c tempest-ServerShowV254Test-42593280 tempest-ServerShowV254Test-42593280-project-member] Lock "b5334832-25dc-4b03-9689-388a1382200a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.945s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.923446] env[63293]: DEBUG nova.policy [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc557354560049119144c62dd61d0816', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2af826d819b4cc4ab27ff8b95144039', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.938164] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528fd2f9-d717-5204-c72f-1e2ce68a5a2b, 'name': SearchDatastore_Task, 'duration_secs': 0.008774} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.938401] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.939017] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] b48324a4-902d-4642-8827-daee9683b3a0/b48324a4-902d-4642-8827-daee9683b3a0.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 803.939017] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f294f1c3-cef6-4fbc-b0db-703cb4414072 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.946148] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for the task: (returnval){ [ 803.946148] env[63293]: value = "task-1327509" [ 803.946148] env[63293]: _type = "Task" [ 803.946148] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.954671] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327509, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.137287] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.137287] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquired lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.137287] env[63293]: DEBUG nova.network.neutron [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 804.146770] env[63293]: DEBUG nova.compute.manager [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Received event network-vif-plugged-70383439-e38d-4f62-8f53-350c7b430342 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.146770] env[63293]: DEBUG oslo_concurrency.lockutils [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] Acquiring lock "495cb1b0-9194-44e6-9f68-de62033ea17d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.146770] env[63293]: DEBUG oslo_concurrency.lockutils [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] Lock "495cb1b0-9194-44e6-9f68-de62033ea17d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.146940] env[63293]: DEBUG oslo_concurrency.lockutils [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] Lock "495cb1b0-9194-44e6-9f68-de62033ea17d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.147035] env[63293]: DEBUG nova.compute.manager [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] No waiting events found dispatching network-vif-plugged-70383439-e38d-4f62-8f53-350c7b430342 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 804.147183] env[63293]: WARNING nova.compute.manager [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Received unexpected event network-vif-plugged-70383439-e38d-4f62-8f53-350c7b430342 for instance with vm_state building and task_state spawning. [ 804.147343] env[63293]: DEBUG nova.compute.manager [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Received event network-changed-70383439-e38d-4f62-8f53-350c7b430342 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.147550] env[63293]: DEBUG nova.compute.manager [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Refreshing instance network info cache due to event network-changed-70383439-e38d-4f62-8f53-350c7b430342. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.147652] env[63293]: DEBUG oslo_concurrency.lockutils [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] Acquiring lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.259175] env[63293]: DEBUG nova.network.neutron [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Successfully created port: 0c416db6-ba22-41a5-b45a-cf9c3f6b1624 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.380770] env[63293]: DEBUG nova.compute.manager [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 804.461504] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327509, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493458} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.464250] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] b48324a4-902d-4642-8827-daee9683b3a0/b48324a4-902d-4642-8827-daee9683b3a0.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 804.464669] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.464975] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4c06a12e-3239-4b82-bd68-05dbacf71006 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.472283] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for the task: (returnval){ [ 804.472283] env[63293]: value = "task-1327510" [ 804.472283] env[63293]: _type = "Task" [ 804.472283] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.486517] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327510, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.718685] env[63293]: DEBUG nova.network.neutron [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.743504] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf78f610-f3db-425d-9b74-254e6f3413df {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.753170] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6abf96-f8df-4e5f-b89a-8c7abfe8e3e6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.796426] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364373c2-b7f1-4d89-9365-f2c4f50f93ab {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.804943] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cbc48a7-2c71-485a-9555-993efb0c6cb3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.818821] env[63293]: DEBUG nova.compute.provider_tree [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.934925] env[63293]: DEBUG nova.network.neutron [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Updating instance_info_cache with network_info: [{"id": "70383439-e38d-4f62-8f53-350c7b430342", "address": "fa:16:3e:ad:36:7b", "network": {"id": "43ed4b72-f94f-47dd-9b14-4297989bd33f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-682241293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3bcac63f7844e10a166eca717f910a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70383439-e3", "ovs_interfaceid": "70383439-e38d-4f62-8f53-350c7b430342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.985830] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327510, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079599} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.986138] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.987022] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d6a711-a7f8-4803-b012-0c298f9d6466 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.017392] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] b48324a4-902d-4642-8827-daee9683b3a0/b48324a4-902d-4642-8827-daee9683b3a0.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.017392] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-167d150e-2d71-48cd-87ff-b6f5cfcf75c5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.037317] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for the task: (returnval){ [ 805.037317] env[63293]: value = "task-1327511" [ 805.037317] env[63293]: _type = "Task" [ 805.037317] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.043559] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327511, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.321700] env[63293]: DEBUG nova.scheduler.client.report [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.396169] env[63293]: DEBUG nova.compute.manager [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 805.427106] env[63293]: DEBUG nova.virt.hardware [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.427507] env[63293]: DEBUG nova.virt.hardware [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.427972] env[63293]: DEBUG nova.virt.hardware [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.428228] env[63293]: DEBUG nova.virt.hardware [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.428493] env[63293]: DEBUG nova.virt.hardware [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.428758] env[63293]: DEBUG nova.virt.hardware [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.429181] env[63293]: DEBUG nova.virt.hardware [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.429407] env[63293]: DEBUG nova.virt.hardware [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.429717] env[63293]: DEBUG nova.virt.hardware [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.430036] env[63293]: DEBUG nova.virt.hardware [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.430609] env[63293]: DEBUG nova.virt.hardware [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.431680] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1e9f43-c853-475a-a622-3bdf84abca33 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.437801] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Releasing lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.438288] env[63293]: DEBUG nova.compute.manager [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Instance network_info: |[{"id": "70383439-e38d-4f62-8f53-350c7b430342", "address": "fa:16:3e:ad:36:7b", "network": {"id": "43ed4b72-f94f-47dd-9b14-4297989bd33f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-682241293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3bcac63f7844e10a166eca717f910a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70383439-e3", "ovs_interfaceid": "70383439-e38d-4f62-8f53-350c7b430342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 805.441728] env[63293]: DEBUG oslo_concurrency.lockutils [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] Acquired lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.442032] env[63293]: DEBUG nova.network.neutron [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Refreshing network info cache for port 70383439-e38d-4f62-8f53-350c7b430342 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 805.443902] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:36:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7e0240aa-a694-48fc-a0f9-6f2d3e71aa12', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '70383439-e38d-4f62-8f53-350c7b430342', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.457756] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Creating folder: Project (a3bcac63f7844e10a166eca717f910a1). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.459835] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f860558-9fb5-43e7-8e68-0e54b054b21e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.464297] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a85b3ff-5717-490b-8dff-2f1a343a7a4b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.491200] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Created folder: Project (a3bcac63f7844e10a166eca717f910a1) in parent group-v283678. [ 805.491200] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Creating folder: Instances. Parent ref: group-v283705. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.491474] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ce8aa7e-0aa7-4117-b316-fa1897d7d34a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.501832] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Created folder: Instances in parent group-v283705. [ 805.504972] env[63293]: DEBUG oslo.service.loopingcall [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.504972] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.504972] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0670745d-0aed-4555-b912-63cb10a92d85 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.540582] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.540582] env[63293]: value = "task-1327514" [ 805.540582] env[63293]: _type = "Task" [ 805.540582] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.551556] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327511, 'name': ReconfigVM_Task, 'duration_secs': 0.344695} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.555060] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Reconfigured VM instance instance-00000036 to attach disk [datastore2] b48324a4-902d-4642-8827-daee9683b3a0/b48324a4-902d-4642-8827-daee9683b3a0.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.555982] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327514, 'name': CreateVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.556294] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7650348a-8058-4d06-a9f2-1d96a6d46922 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.565181] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for the task: (returnval){ [ 805.565181] env[63293]: value = "task-1327515" [ 805.565181] env[63293]: _type = "Task" [ 805.565181] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.575237] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327515, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.827033] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.827394] env[63293]: DEBUG nova.compute.manager [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 805.830282] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.434s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.832248] env[63293]: INFO nova.compute.claims [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.053309] env[63293]: DEBUG nova.network.neutron [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Successfully updated port: 0c416db6-ba22-41a5-b45a-cf9c3f6b1624 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.060988] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327514, 'name': CreateVM_Task, 'duration_secs': 0.425076} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.064014] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 806.064767] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.065877] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.065877] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 806.065992] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acf46259-cc2b-4825-a282-e6f1e5f0e78a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.074759] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for the task: (returnval){ [ 806.074759] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f1477d-7298-9204-550a-c3d0e6d430bf" [ 806.074759] env[63293]: _type = "Task" [ 806.074759] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.078158] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327515, 'name': Rename_Task, 'duration_secs': 0.225747} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.080867] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 806.081140] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d72814b8-3f8a-4b6e-8bc8-0ecb9332c123 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.091489] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f1477d-7298-9204-550a-c3d0e6d430bf, 'name': SearchDatastore_Task, 'duration_secs': 0.009284} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.092826] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.093079] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.093310] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.093456] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.093630] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.093927] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for the task: (returnval){ [ 806.093927] env[63293]: value = "task-1327516" [ 806.093927] env[63293]: _type = "Task" [ 806.093927] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.095717] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0be245e5-31d9-4b42-bf4c-083762a43ed4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.105786] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327516, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.109171] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.109171] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 806.110094] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2adfce5b-eef9-4c0e-8b22-960bc13371f3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.115613] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for the task: (returnval){ [ 806.115613] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524d082b-7430-4025-5882-2d89072ec631" [ 806.115613] env[63293]: _type = "Task" [ 806.115613] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.124067] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524d082b-7430-4025-5882-2d89072ec631, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.226031] env[63293]: DEBUG nova.compute.manager [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Received event network-vif-plugged-0c416db6-ba22-41a5-b45a-cf9c3f6b1624 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.226270] env[63293]: DEBUG oslo_concurrency.lockutils [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] Acquiring lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.226336] env[63293]: DEBUG oslo_concurrency.lockutils [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] Lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.226499] env[63293]: DEBUG oslo_concurrency.lockutils [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] Lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.226672] env[63293]: DEBUG nova.compute.manager [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] No waiting events found dispatching network-vif-plugged-0c416db6-ba22-41a5-b45a-cf9c3f6b1624 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 806.226861] env[63293]: WARNING nova.compute.manager [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Received unexpected event network-vif-plugged-0c416db6-ba22-41a5-b45a-cf9c3f6b1624 for instance with vm_state building and task_state spawning. [ 806.227042] env[63293]: DEBUG nova.compute.manager [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Received event network-changed-0c416db6-ba22-41a5-b45a-cf9c3f6b1624 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.227200] env[63293]: DEBUG nova.compute.manager [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Refreshing instance network info cache due to event network-changed-0c416db6-ba22-41a5-b45a-cf9c3f6b1624. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 806.227379] env[63293]: DEBUG oslo_concurrency.lockutils [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] Acquiring lock "refresh_cache-cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.227513] env[63293]: DEBUG oslo_concurrency.lockutils [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] Acquired lock "refresh_cache-cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.227666] env[63293]: DEBUG nova.network.neutron [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Refreshing network info cache for port 0c416db6-ba22-41a5-b45a-cf9c3f6b1624 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 806.248164] env[63293]: DEBUG nova.network.neutron [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Updated VIF entry in instance network info cache for port 70383439-e38d-4f62-8f53-350c7b430342. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 806.248815] env[63293]: DEBUG nova.network.neutron [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Updating instance_info_cache with network_info: [{"id": "70383439-e38d-4f62-8f53-350c7b430342", "address": "fa:16:3e:ad:36:7b", "network": {"id": "43ed4b72-f94f-47dd-9b14-4297989bd33f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-682241293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3bcac63f7844e10a166eca717f910a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70383439-e3", "ovs_interfaceid": "70383439-e38d-4f62-8f53-350c7b430342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.337304] env[63293]: DEBUG nova.compute.utils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 806.340449] env[63293]: DEBUG nova.compute.manager [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 806.340658] env[63293]: DEBUG nova.network.neutron [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 806.381204] env[63293]: DEBUG nova.policy [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5dab7cefbc2c4360b927f4c7e5e5319e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e1a30d37c1b47fd8670db0a9ef7020d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 806.562536] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "refresh_cache-cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.611444] env[63293]: DEBUG oslo_vmware.api [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327516, 'name': PowerOnVM_Task, 'duration_secs': 0.499811} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.611444] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 806.611623] env[63293]: INFO nova.compute.manager [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Took 7.08 seconds to spawn the instance on the hypervisor. [ 806.613746] env[63293]: DEBUG nova.compute.manager [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.613746] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d294e3e-ea1c-46b0-b1ea-d012a4a8a036 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.633994] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524d082b-7430-4025-5882-2d89072ec631, 'name': SearchDatastore_Task, 'duration_secs': 0.009058} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.635302] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74819af4-e4d0-446a-a765-0dca2fa1f6ca {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.643328] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for the task: (returnval){ [ 806.643328] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525e3008-5d2f-2987-cf3b-eed3b9edb2da" [ 806.643328] env[63293]: _type = "Task" [ 806.643328] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.653559] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525e3008-5d2f-2987-cf3b-eed3b9edb2da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.661997] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "d753598d-a92a-4515-9ad1-d386294f7a99" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.662249] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.751631] env[63293]: DEBUG oslo_concurrency.lockutils [req-489ebf13-1949-4170-8faf-f46bc0273323 req-50068af6-afaf-450e-8540-90500b000b70 service nova] Releasing lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.760562] env[63293]: DEBUG nova.network.neutron [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.805107] env[63293]: DEBUG nova.network.neutron [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Successfully created port: 2f0bfb1c-6184-49f1-9c71-cb7459e1470d {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 806.844204] env[63293]: DEBUG nova.compute.manager [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 806.855020] env[63293]: DEBUG nova.network.neutron [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.134505] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18014ed3-8765-4cb3-8715-fd9523feba18 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.145491] env[63293]: INFO nova.compute.manager [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Took 40.55 seconds to build instance. [ 807.150287] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404f6fa5-a298-4d1b-aaa3-672d369946bc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.183187] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525e3008-5d2f-2987-cf3b-eed3b9edb2da, 'name': SearchDatastore_Task, 'duration_secs': 0.009263} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.184474] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ff1684-c081-4203-8b9b-b49995c8d2fb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.186985] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.187263] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 495cb1b0-9194-44e6-9f68-de62033ea17d/495cb1b0-9194-44e6-9f68-de62033ea17d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 807.187738] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-baef0838-fbc1-40e6-a304-79f94cc1690e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.195898] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef346ca0-6287-4f83-b59a-7ca7e62499b9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.200817] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for the task: (returnval){ [ 807.200817] env[63293]: value = "task-1327517" [ 807.200817] env[63293]: _type = "Task" [ 807.200817] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.213696] env[63293]: DEBUG nova.compute.provider_tree [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.221155] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327517, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.357474] env[63293]: DEBUG oslo_concurrency.lockutils [req-84d9ccb5-def0-4356-8ad7-fe4440436d88 req-62c20cbe-6577-423f-b1c8-eca3da408bb8 service nova] Releasing lock "refresh_cache-cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.357628] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "refresh_cache-cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.357934] env[63293]: DEBUG nova.network.neutron [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.655012] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5441c2c1-81a8-45e9-b42f-97d37dfba582 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "b48324a4-902d-4642-8827-daee9683b3a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.268s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.711884] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327517, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473423} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.712168] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 495cb1b0-9194-44e6-9f68-de62033ea17d/495cb1b0-9194-44e6-9f68-de62033ea17d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 807.712382] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.712626] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7fa7689a-70cd-4b91-a86c-5222ab70f561 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.716471] env[63293]: DEBUG nova.scheduler.client.report [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.721057] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for the task: (returnval){ [ 807.721057] env[63293]: value = "task-1327518" [ 807.721057] env[63293]: _type = "Task" [ 807.721057] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.728994] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327518, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.859033] env[63293]: DEBUG nova.compute.manager [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 807.884210] env[63293]: DEBUG nova.virt.hardware [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 807.884496] env[63293]: DEBUG nova.virt.hardware [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 807.884675] env[63293]: DEBUG nova.virt.hardware [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.884895] env[63293]: DEBUG nova.virt.hardware [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 807.885089] env[63293]: DEBUG nova.virt.hardware [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.885318] env[63293]: DEBUG nova.virt.hardware [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 807.885654] env[63293]: DEBUG nova.virt.hardware [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 807.885906] env[63293]: DEBUG nova.virt.hardware [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 807.886187] env[63293]: DEBUG nova.virt.hardware [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 807.886452] env[63293]: DEBUG nova.virt.hardware [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 807.886728] env[63293]: DEBUG nova.virt.hardware [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.887702] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37a56e4-8b14-4d76-9e3d-46dcd2e2a4fd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.891923] env[63293]: DEBUG nova.network.neutron [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.897956] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8587f1-0766-435b-a7ba-f15214d8a004 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.026830] env[63293]: DEBUG nova.network.neutron [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Updating instance_info_cache with network_info: [{"id": "0c416db6-ba22-41a5-b45a-cf9c3f6b1624", "address": "fa:16:3e:7d:90:49", "network": {"id": "340bfd3b-7dff-454b-87df-788003988adc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-388838680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2af826d819b4cc4ab27ff8b95144039", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c416db6-ba", "ovs_interfaceid": "0c416db6-ba22-41a5-b45a-cf9c3f6b1624", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.157285] env[63293]: DEBUG nova.compute.manager [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 808.224883] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.224883] env[63293]: DEBUG nova.compute.manager [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 808.225885] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.941s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.230503] env[63293]: INFO nova.compute.claims [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 808.241180] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327518, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097168} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.241578] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.242489] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd2c679-c91a-478d-96fe-a3f24f8bf73b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.271772] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 495cb1b0-9194-44e6-9f68-de62033ea17d/495cb1b0-9194-44e6-9f68-de62033ea17d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.272948] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ba545de-39a2-4a45-8503-f46558525434 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.297018] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for the task: (returnval){ [ 808.297018] env[63293]: value = "task-1327519" [ 808.297018] env[63293]: _type = "Task" [ 808.297018] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.303493] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327519, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.328365] env[63293]: DEBUG nova.compute.manager [req-8cce004e-635c-4e95-a4fa-9a72e6f76379 req-93ed5ba6-12e1-45b1-8a25-7b22cc5db540 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Received event network-vif-plugged-2f0bfb1c-6184-49f1-9c71-cb7459e1470d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.328698] env[63293]: DEBUG oslo_concurrency.lockutils [req-8cce004e-635c-4e95-a4fa-9a72e6f76379 req-93ed5ba6-12e1-45b1-8a25-7b22cc5db540 service nova] Acquiring lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.330066] env[63293]: DEBUG oslo_concurrency.lockutils [req-8cce004e-635c-4e95-a4fa-9a72e6f76379 req-93ed5ba6-12e1-45b1-8a25-7b22cc5db540 service nova] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.330066] env[63293]: DEBUG oslo_concurrency.lockutils [req-8cce004e-635c-4e95-a4fa-9a72e6f76379 req-93ed5ba6-12e1-45b1-8a25-7b22cc5db540 service nova] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.330066] env[63293]: DEBUG nova.compute.manager [req-8cce004e-635c-4e95-a4fa-9a72e6f76379 req-93ed5ba6-12e1-45b1-8a25-7b22cc5db540 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] No waiting events found dispatching network-vif-plugged-2f0bfb1c-6184-49f1-9c71-cb7459e1470d {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 808.330066] env[63293]: WARNING nova.compute.manager [req-8cce004e-635c-4e95-a4fa-9a72e6f76379 req-93ed5ba6-12e1-45b1-8a25-7b22cc5db540 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Received unexpected event network-vif-plugged-2f0bfb1c-6184-49f1-9c71-cb7459e1470d for instance with vm_state building and task_state spawning. [ 808.431837] env[63293]: DEBUG nova.network.neutron [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Successfully updated port: 2f0bfb1c-6184-49f1-9c71-cb7459e1470d {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 808.529745] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "refresh_cache-cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.530232] env[63293]: DEBUG nova.compute.manager [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Instance network_info: |[{"id": "0c416db6-ba22-41a5-b45a-cf9c3f6b1624", "address": "fa:16:3e:7d:90:49", "network": {"id": "340bfd3b-7dff-454b-87df-788003988adc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-388838680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2af826d819b4cc4ab27ff8b95144039", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c416db6-ba", "ovs_interfaceid": "0c416db6-ba22-41a5-b45a-cf9c3f6b1624", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 808.530705] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:90:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb2eaecd-9701-4504-9fcb-fb1a420ead72', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c416db6-ba22-41a5-b45a-cf9c3f6b1624', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.538322] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Creating folder: Project (a2af826d819b4cc4ab27ff8b95144039). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 808.538760] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-400c44dd-8997-4c99-9476-f3f374b0ea2d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.549440] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Created folder: Project (a2af826d819b4cc4ab27ff8b95144039) in parent group-v283678. [ 808.549591] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Creating folder: Instances. Parent ref: group-v283708. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 808.549824] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7df09d30-4b09-4403-a99e-7d1e566a72d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.559125] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Created folder: Instances in parent group-v283708. [ 808.559411] env[63293]: DEBUG oslo.service.loopingcall [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.559536] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 808.559738] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac02877b-6891-4d77-bd53-1cb2a10290e1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.578182] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.578182] env[63293]: value = "task-1327522" [ 808.578182] env[63293]: _type = "Task" [ 808.578182] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.585496] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327522, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.681505] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.738773] env[63293]: DEBUG nova.compute.utils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 808.740135] env[63293]: DEBUG nova.compute.manager [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 808.740307] env[63293]: DEBUG nova.network.neutron [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 808.805530] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327519, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.813548] env[63293]: DEBUG nova.policy [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4f532f456c154bb79945183ebc2a9a49', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f36b230cea2e491ba6c812c66e5ada66', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 808.934855] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.934855] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquired lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.934855] env[63293]: DEBUG nova.network.neutron [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.088961] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327522, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.194790] env[63293]: DEBUG nova.network.neutron [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Successfully created port: e35e7768-ec67-4322-a4c5-be6a5835054e {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 809.244489] env[63293]: DEBUG nova.compute.manager [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 809.306386] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327519, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.478983] env[63293]: DEBUG nova.network.neutron [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.586993] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41760175-8a24-4d44-943d-c43666532046 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.597142] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5aec02-154b-493e-8479-3d378a85f0f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.600399] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327522, 'name': CreateVM_Task, 'duration_secs': 0.719271} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.602529] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 809.604222] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.604222] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.604222] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.604841] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c7005db-c912-4eb7-9884-b7998e646df4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.630227] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bc202d-a463-4e27-80be-0bed1e4b0be2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.633975] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 809.633975] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ace67f-62a7-fb93-8f7e-f99642b0ca5b" [ 809.633975] env[63293]: _type = "Task" [ 809.633975] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.640475] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cafafd6-59be-4891-a396-48223dcf73d2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.646573] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ace67f-62a7-fb93-8f7e-f99642b0ca5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.655868] env[63293]: DEBUG nova.compute.provider_tree [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.657778] env[63293]: DEBUG nova.network.neutron [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updating instance_info_cache with network_info: [{"id": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "address": "fa:16:3e:80:11:9c", "network": {"id": "9d5a3e58-567a-4c3f-9ad9-9fcbf81fced9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-949261114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1a30d37c1b47fd8670db0a9ef7020d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f0bfb1c-61", "ovs_interfaceid": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.806039] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327519, 'name': ReconfigVM_Task, 'duration_secs': 1.299692} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.806297] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 495cb1b0-9194-44e6-9f68-de62033ea17d/495cb1b0-9194-44e6-9f68-de62033ea17d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.806949] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0fe98241-d1af-4734-8fd6-75e3fd8ae010 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.813270] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for the task: (returnval){ [ 809.813270] env[63293]: value = "task-1327523" [ 809.813270] env[63293]: _type = "Task" [ 809.813270] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.825738] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327523, 'name': Rename_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.143982] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ace67f-62a7-fb93-8f7e-f99642b0ca5b, 'name': SearchDatastore_Task, 'duration_secs': 0.010809} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.144300] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.144523] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 810.144754] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.144898] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.145086] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 810.145337] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0609626b-e9d4-4994-9cd4-fd576552d04e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.153358] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 810.153572] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 810.154292] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e16b3446-52c8-423f-84ea-86da9b6a6090 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.160103] env[63293]: DEBUG nova.scheduler.client.report [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.163465] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Releasing lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.163716] env[63293]: DEBUG nova.compute.manager [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Instance network_info: |[{"id": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "address": "fa:16:3e:80:11:9c", "network": {"id": "9d5a3e58-567a-4c3f-9ad9-9fcbf81fced9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-949261114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1a30d37c1b47fd8670db0a9ef7020d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f0bfb1c-61", "ovs_interfaceid": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 810.164027] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 810.164027] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52dfa78d-37f9-3c9e-5f27-fac9bcd067db" [ 810.164027] env[63293]: _type = "Task" [ 810.164027] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.164552] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:11:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2f0bfb1c-6184-49f1-9c71-cb7459e1470d', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 810.172185] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Creating folder: Project (3e1a30d37c1b47fd8670db0a9ef7020d). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 810.172506] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a95f7872-42d2-4f43-8583-15c36653ccfc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.183271] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52dfa78d-37f9-3c9e-5f27-fac9bcd067db, 'name': SearchDatastore_Task, 'duration_secs': 0.008517} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.184865] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ce8843e-8a23-4048-a44f-3fcf9651db13 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.186844] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Created folder: Project (3e1a30d37c1b47fd8670db0a9ef7020d) in parent group-v283678. [ 810.187034] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Creating folder: Instances. Parent ref: group-v283711. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 810.187505] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-20b6bedc-a826-47db-a992-1a1dc362868b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.191568] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 810.191568] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527450b8-3a78-46d0-0622-156bdc64e581" [ 810.191568] env[63293]: _type = "Task" [ 810.191568] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.196699] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Created folder: Instances in parent group-v283711. [ 810.196956] env[63293]: DEBUG oslo.service.loopingcall [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.199856] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 810.200107] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527450b8-3a78-46d0-0622-156bdc64e581, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.200291] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2238e9b-27fb-44f1-bd5a-85e049ce7871 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.218871] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 810.218871] env[63293]: value = "task-1327526" [ 810.218871] env[63293]: _type = "Task" [ 810.218871] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.225757] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327526, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.253678] env[63293]: DEBUG nova.compute.manager [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 810.279103] env[63293]: DEBUG nova.virt.hardware [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 810.279353] env[63293]: DEBUG nova.virt.hardware [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 810.279508] env[63293]: DEBUG nova.virt.hardware [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.279691] env[63293]: DEBUG nova.virt.hardware [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 810.279833] env[63293]: DEBUG nova.virt.hardware [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.280071] env[63293]: DEBUG nova.virt.hardware [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 810.280289] env[63293]: DEBUG nova.virt.hardware [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 810.280464] env[63293]: DEBUG nova.virt.hardware [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 810.280634] env[63293]: DEBUG nova.virt.hardware [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 810.280793] env[63293]: DEBUG nova.virt.hardware [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 810.280963] env[63293]: DEBUG nova.virt.hardware [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 810.281836] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9fe798-bdfa-48f0-9189-17ced83212f0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.290875] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebe2f46-2c79-4119-81f2-88695b62357c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.321670] env[63293]: DEBUG oslo_concurrency.lockutils [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "b48324a4-902d-4642-8827-daee9683b3a0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.321897] env[63293]: DEBUG oslo_concurrency.lockutils [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "b48324a4-902d-4642-8827-daee9683b3a0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.322127] env[63293]: DEBUG oslo_concurrency.lockutils [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "b48324a4-902d-4642-8827-daee9683b3a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.322333] env[63293]: DEBUG oslo_concurrency.lockutils [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "b48324a4-902d-4642-8827-daee9683b3a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.323034] env[63293]: DEBUG oslo_concurrency.lockutils [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "b48324a4-902d-4642-8827-daee9683b3a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.324200] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327523, 'name': Rename_Task, 'duration_secs': 0.150068} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.324591] env[63293]: INFO nova.compute.manager [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Terminating instance [ 810.325897] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.326445] env[63293]: DEBUG nova.compute.manager [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 810.326629] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 810.326851] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb5d97af-b3d6-42a2-9322-417c8d3b36ef {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.329082] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24ab5db-978a-4f54-b73a-60dcc652db35 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.336199] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 810.337391] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d63b6e0b-a085-46df-8eef-585158d79b5e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.338813] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for the task: (returnval){ [ 810.338813] env[63293]: value = "task-1327527" [ 810.338813] env[63293]: _type = "Task" [ 810.338813] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.344048] env[63293]: DEBUG oslo_vmware.api [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for the task: (returnval){ [ 810.344048] env[63293]: value = "task-1327528" [ 810.344048] env[63293]: _type = "Task" [ 810.344048] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.347021] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327527, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.352881] env[63293]: DEBUG nova.compute.manager [req-569a405c-5f22-409d-81ec-eee596f7c716 req-ea3c4cce-edaf-4605-8ead-0513f42650d1 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Received event network-changed-2f0bfb1c-6184-49f1-9c71-cb7459e1470d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.353068] env[63293]: DEBUG nova.compute.manager [req-569a405c-5f22-409d-81ec-eee596f7c716 req-ea3c4cce-edaf-4605-8ead-0513f42650d1 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Refreshing instance network info cache due to event network-changed-2f0bfb1c-6184-49f1-9c71-cb7459e1470d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 810.353164] env[63293]: DEBUG oslo_concurrency.lockutils [req-569a405c-5f22-409d-81ec-eee596f7c716 req-ea3c4cce-edaf-4605-8ead-0513f42650d1 service nova] Acquiring lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.353351] env[63293]: DEBUG oslo_concurrency.lockutils [req-569a405c-5f22-409d-81ec-eee596f7c716 req-ea3c4cce-edaf-4605-8ead-0513f42650d1 service nova] Acquired lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.353489] env[63293]: DEBUG nova.network.neutron [req-569a405c-5f22-409d-81ec-eee596f7c716 req-ea3c4cce-edaf-4605-8ead-0513f42650d1 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Refreshing network info cache for port 2f0bfb1c-6184-49f1-9c71-cb7459e1470d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 810.359119] env[63293]: DEBUG oslo_vmware.api [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327528, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.665701] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.666240] env[63293]: DEBUG nova.compute.manager [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 810.669189] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.413s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.670566] env[63293]: INFO nova.compute.claims [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.702357] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527450b8-3a78-46d0-0622-156bdc64e581, 'name': SearchDatastore_Task, 'duration_secs': 0.009443} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.702357] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.702739] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395/cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.702875] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3434eee6-cb81-45c5-9daa-509889f3f73a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.710400] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 810.710400] env[63293]: value = "task-1327529" [ 810.710400] env[63293]: _type = "Task" [ 810.710400] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.719375] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327529, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.726928] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327526, 'name': CreateVM_Task, 'duration_secs': 0.361713} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.727603] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 810.727958] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.728163] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.728644] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 810.728720] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f50a8a1-9f36-493e-8504-30a65cace26d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.733505] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 810.733505] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52883d9e-3009-f78e-1702-120b562d7362" [ 810.733505] env[63293]: _type = "Task" [ 810.733505] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.741440] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52883d9e-3009-f78e-1702-120b562d7362, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.812265] env[63293]: DEBUG nova.compute.manager [req-66430b05-6d98-4114-9ca3-eceee3cee74a req-9254ba9f-4cff-4b09-b0d2-40e17945aadb service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Received event network-vif-plugged-e35e7768-ec67-4322-a4c5-be6a5835054e {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.812509] env[63293]: DEBUG oslo_concurrency.lockutils [req-66430b05-6d98-4114-9ca3-eceee3cee74a req-9254ba9f-4cff-4b09-b0d2-40e17945aadb service nova] Acquiring lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.812737] env[63293]: DEBUG oslo_concurrency.lockutils [req-66430b05-6d98-4114-9ca3-eceee3cee74a req-9254ba9f-4cff-4b09-b0d2-40e17945aadb service nova] Lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.812836] env[63293]: DEBUG oslo_concurrency.lockutils [req-66430b05-6d98-4114-9ca3-eceee3cee74a req-9254ba9f-4cff-4b09-b0d2-40e17945aadb service nova] Lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.813122] env[63293]: DEBUG nova.compute.manager [req-66430b05-6d98-4114-9ca3-eceee3cee74a req-9254ba9f-4cff-4b09-b0d2-40e17945aadb service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] No waiting events found dispatching network-vif-plugged-e35e7768-ec67-4322-a4c5-be6a5835054e {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 810.813471] env[63293]: WARNING nova.compute.manager [req-66430b05-6d98-4114-9ca3-eceee3cee74a req-9254ba9f-4cff-4b09-b0d2-40e17945aadb service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Received unexpected event network-vif-plugged-e35e7768-ec67-4322-a4c5-be6a5835054e for instance with vm_state building and task_state spawning. [ 810.849650] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327527, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.863103] env[63293]: DEBUG oslo_vmware.api [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327528, 'name': PowerOffVM_Task, 'duration_secs': 0.22241} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.863494] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 810.864147] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 810.864466] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-012f6f96-54d9-4cb0-8e73-084b8ff72ed3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.930920] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 810.931693] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 810.931693] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Deleting the datastore file [datastore2] b48324a4-902d-4642-8827-daee9683b3a0 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.931693] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cd80f552-5e42-4135-a863-3925e47571ac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.940911] env[63293]: DEBUG oslo_vmware.api [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for the task: (returnval){ [ 810.940911] env[63293]: value = "task-1327531" [ 810.940911] env[63293]: _type = "Task" [ 810.940911] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.951942] env[63293]: DEBUG oslo_vmware.api [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.175620] env[63293]: DEBUG nova.compute.utils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 811.184533] env[63293]: DEBUG nova.network.neutron [req-569a405c-5f22-409d-81ec-eee596f7c716 req-ea3c4cce-edaf-4605-8ead-0513f42650d1 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updated VIF entry in instance network info cache for port 2f0bfb1c-6184-49f1-9c71-cb7459e1470d. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 811.184976] env[63293]: DEBUG nova.network.neutron [req-569a405c-5f22-409d-81ec-eee596f7c716 req-ea3c4cce-edaf-4605-8ead-0513f42650d1 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updating instance_info_cache with network_info: [{"id": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "address": "fa:16:3e:80:11:9c", "network": {"id": "9d5a3e58-567a-4c3f-9ad9-9fcbf81fced9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-949261114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1a30d37c1b47fd8670db0a9ef7020d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f0bfb1c-61", "ovs_interfaceid": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.189759] env[63293]: DEBUG nova.compute.manager [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 811.189996] env[63293]: DEBUG nova.network.neutron [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 811.220557] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327529, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504262} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.221116] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395/cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.221116] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.221383] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f56543cc-e16a-47f6-a324-a5a5b8f29ddf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.227410] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 811.227410] env[63293]: value = "task-1327532" [ 811.227410] env[63293]: _type = "Task" [ 811.227410] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.240576] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327532, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.247080] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52883d9e-3009-f78e-1702-120b562d7362, 'name': SearchDatastore_Task, 'duration_secs': 0.008679} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.247402] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.247665] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.247969] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.248105] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.248308] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.248596] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04d2a749-123d-4e98-a2b4-c7e5f59df15e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.254600] env[63293]: DEBUG nova.policy [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2312d48038d7474481a0709267645780', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '866b347100454019b07f63922b995bb6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 811.257594] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.257693] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 811.258572] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-019105d5-37bc-4ef1-bf91-916f28708ecf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.264713] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 811.264713] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ebe25f-d18a-255f-4949-671771fc6773" [ 811.264713] env[63293]: _type = "Task" [ 811.264713] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.276453] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ebe25f-d18a-255f-4949-671771fc6773, 'name': SearchDatastore_Task, 'duration_secs': 0.00911} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.277391] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d6e067f-7c40-4371-95e8-c80c0ce50f34 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.283166] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 811.283166] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]521d2e72-c993-a6e8-093e-0f95f29c1732" [ 811.283166] env[63293]: _type = "Task" [ 811.283166] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.293280] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]521d2e72-c993-a6e8-093e-0f95f29c1732, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.341447] env[63293]: DEBUG nova.network.neutron [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Successfully updated port: e35e7768-ec67-4322-a4c5-be6a5835054e {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 811.355500] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327527, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.451552] env[63293]: DEBUG oslo_vmware.api [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Task: {'id': task-1327531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.374363} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.451814] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 811.452121] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 811.452359] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 811.452549] env[63293]: INFO nova.compute.manager [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Took 1.13 seconds to destroy the instance on the hypervisor. [ 811.456018] env[63293]: DEBUG oslo.service.loopingcall [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.456018] env[63293]: DEBUG nova.compute.manager [-] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 811.456018] env[63293]: DEBUG nova.network.neutron [-] [instance: b48324a4-902d-4642-8827-daee9683b3a0] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 811.692946] env[63293]: DEBUG nova.network.neutron [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Successfully created port: a34b2cc4-afe1-463c-bb22-470ea54d8789 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 811.697241] env[63293]: DEBUG oslo_concurrency.lockutils [req-569a405c-5f22-409d-81ec-eee596f7c716 req-ea3c4cce-edaf-4605-8ead-0513f42650d1 service nova] Releasing lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.701914] env[63293]: DEBUG nova.compute.manager [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 811.742847] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327532, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067379} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.743158] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.743975] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95533b3c-5a83-4d13-baa8-c13f6fa2b48d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.778314] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395/cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.782987] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b6f7335-4b6e-4a06-b156-b03720a6c0e5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.807261] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]521d2e72-c993-a6e8-093e-0f95f29c1732, 'name': SearchDatastore_Task, 'duration_secs': 0.010523} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.808512] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.808779] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 0977c79a-2ef3-4ec9-a0a0-de1ed5799723/0977c79a-2ef3-4ec9-a0a0-de1ed5799723.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 811.809133] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 811.809133] env[63293]: value = "task-1327533" [ 811.809133] env[63293]: _type = "Task" [ 811.809133] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.811534] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-82b25e6e-70b7-41f8-8c50-b21b59a131c8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.822598] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327533, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.823469] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 811.823469] env[63293]: value = "task-1327534" [ 811.823469] env[63293]: _type = "Task" [ 811.823469] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.831375] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327534, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.851458] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquiring lock "refresh_cache-c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.851458] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquired lock "refresh_cache-c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.851458] env[63293]: DEBUG nova.network.neutron [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 811.856235] env[63293]: DEBUG oslo_vmware.api [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327527, 'name': PowerOnVM_Task, 'duration_secs': 1.410518} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.856459] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 811.856690] env[63293]: INFO nova.compute.manager [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Took 8.90 seconds to spawn the instance on the hypervisor. [ 811.856950] env[63293]: DEBUG nova.compute.manager [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.857857] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2af516b-10b2-488f-b640-1f087ef2c07e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.070168] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d25973-3d61-493b-a2de-2b7267bf43d8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.081739] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c283069-a959-43e0-b0c8-5a42de5b15c1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.117985] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2125bb-7e4b-45b7-aa15-ca35ce24d8a1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.129227] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb15a4e-a01c-4489-8c04-89fe620b1d3f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.147610] env[63293]: DEBUG nova.compute.provider_tree [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.298988] env[63293]: DEBUG nova.network.neutron [-] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.324327] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327533, 'name': ReconfigVM_Task, 'duration_secs': 0.27475} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.324428] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Reconfigured VM instance instance-00000038 to attach disk [datastore2] cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395/cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.327784] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8dceec11-f85c-48c7-ac63-55cc55134d0e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.335248] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327534, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486799} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.336517] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 0977c79a-2ef3-4ec9-a0a0-de1ed5799723/0977c79a-2ef3-4ec9-a0a0-de1ed5799723.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.336789] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.336992] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 812.336992] env[63293]: value = "task-1327535" [ 812.336992] env[63293]: _type = "Task" [ 812.336992] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.337187] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f92ad8f-6846-4f2b-8db3-36d1d478fc99 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.346961] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327535, 'name': Rename_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.347944] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 812.347944] env[63293]: value = "task-1327536" [ 812.347944] env[63293]: _type = "Task" [ 812.347944] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.359503] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327536, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.373182] env[63293]: INFO nova.compute.manager [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Took 41.90 seconds to build instance. [ 812.401750] env[63293]: DEBUG nova.network.neutron [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.425227] env[63293]: DEBUG nova.compute.manager [req-40d0d556-0635-4234-86f3-7bc67bf74a5e req-1b368108-78e7-4d34-82ac-a4994d85ca68 service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Received event network-changed-e35e7768-ec67-4322-a4c5-be6a5835054e {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.425511] env[63293]: DEBUG nova.compute.manager [req-40d0d556-0635-4234-86f3-7bc67bf74a5e req-1b368108-78e7-4d34-82ac-a4994d85ca68 service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Refreshing instance network info cache due to event network-changed-e35e7768-ec67-4322-a4c5-be6a5835054e. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 812.425752] env[63293]: DEBUG oslo_concurrency.lockutils [req-40d0d556-0635-4234-86f3-7bc67bf74a5e req-1b368108-78e7-4d34-82ac-a4994d85ca68 service nova] Acquiring lock "refresh_cache-c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.594111] env[63293]: DEBUG nova.network.neutron [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Updating instance_info_cache with network_info: [{"id": "e35e7768-ec67-4322-a4c5-be6a5835054e", "address": "fa:16:3e:63:30:1e", "network": {"id": "02c216f2-c41a-4722-91df-330cd520be3d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.39", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f57886847a1446109c6ab061a248dcf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35e7768-ec", "ovs_interfaceid": "e35e7768-ec67-4322-a4c5-be6a5835054e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.653022] env[63293]: DEBUG nova.scheduler.client.report [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.713591] env[63293]: DEBUG nova.compute.manager [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 812.738062] env[63293]: DEBUG nova.virt.hardware [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 812.738325] env[63293]: DEBUG nova.virt.hardware [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 812.738483] env[63293]: DEBUG nova.virt.hardware [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 812.738666] env[63293]: DEBUG nova.virt.hardware [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 812.738834] env[63293]: DEBUG nova.virt.hardware [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 812.739016] env[63293]: DEBUG nova.virt.hardware [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 812.739273] env[63293]: DEBUG nova.virt.hardware [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 812.739460] env[63293]: DEBUG nova.virt.hardware [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 812.739652] env[63293]: DEBUG nova.virt.hardware [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 812.739850] env[63293]: DEBUG nova.virt.hardware [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 812.740105] env[63293]: DEBUG nova.virt.hardware [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 812.741316] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4c45ed-2a2a-4999-aa4f-24dd28d70ed0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.749399] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f55b44-1189-4bb6-9d03-fafe93e1e5b3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.801580] env[63293]: INFO nova.compute.manager [-] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Took 1.35 seconds to deallocate network for instance. [ 812.849503] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327535, 'name': Rename_Task, 'duration_secs': 0.146526} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.852595] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 812.852840] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d380a937-c3c8-4771-926b-8b96e61987e6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.860039] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327536, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.204439} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.861194] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.861517] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 812.861517] env[63293]: value = "task-1327537" [ 812.861517] env[63293]: _type = "Task" [ 812.861517] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.862188] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4dc4558-fce6-461f-aec8-ea77f988e885 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.871659] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327537, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.881208] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d7f950e7-c0fc-4a9e-a1db-532123f4412f tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "495cb1b0-9194-44e6-9f68-de62033ea17d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.998s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.890359] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 0977c79a-2ef3-4ec9-a0a0-de1ed5799723/0977c79a-2ef3-4ec9-a0a0-de1ed5799723.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.891623] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eadf37d1-b48d-41c3-a770-d13c0179723b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.913542] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 812.913542] env[63293]: value = "task-1327538" [ 812.913542] env[63293]: _type = "Task" [ 812.913542] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.921308] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327538, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.097252] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Releasing lock "refresh_cache-c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.097655] env[63293]: DEBUG nova.compute.manager [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Instance network_info: |[{"id": "e35e7768-ec67-4322-a4c5-be6a5835054e", "address": "fa:16:3e:63:30:1e", "network": {"id": "02c216f2-c41a-4722-91df-330cd520be3d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.39", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f57886847a1446109c6ab061a248dcf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35e7768-ec", "ovs_interfaceid": "e35e7768-ec67-4322-a4c5-be6a5835054e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 813.097977] env[63293]: DEBUG oslo_concurrency.lockutils [req-40d0d556-0635-4234-86f3-7bc67bf74a5e req-1b368108-78e7-4d34-82ac-a4994d85ca68 service nova] Acquired lock "refresh_cache-c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.098174] env[63293]: DEBUG nova.network.neutron [req-40d0d556-0635-4234-86f3-7bc67bf74a5e req-1b368108-78e7-4d34-82ac-a4994d85ca68 service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Refreshing network info cache for port e35e7768-ec67-4322-a4c5-be6a5835054e {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 813.099563] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:30:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e35e7768-ec67-4322-a4c5-be6a5835054e', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.107439] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Creating folder: Project (f36b230cea2e491ba6c812c66e5ada66). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 813.110867] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96522e6a-893d-4bb3-b4f1-80270eba8698 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.122165] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Created folder: Project (f36b230cea2e491ba6c812c66e5ada66) in parent group-v283678. [ 813.122428] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Creating folder: Instances. Parent ref: group-v283714. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 813.122703] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-95e4aab8-e10a-4491-b473-65049fa7276b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.132165] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Created folder: Instances in parent group-v283714. [ 813.132399] env[63293]: DEBUG oslo.service.loopingcall [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.132594] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 813.132826] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-72168736-4fd4-4105-aa47-ce943bd14a79 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.152870] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.152870] env[63293]: value = "task-1327541" [ 813.152870] env[63293]: _type = "Task" [ 813.152870] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.157061] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.157536] env[63293]: DEBUG nova.compute.manager [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 813.162090] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.722s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.171585] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327541, 'name': CreateVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.308499] env[63293]: DEBUG oslo_concurrency.lockutils [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.376066] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327537, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.409144] env[63293]: DEBUG nova.compute.manager [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.420372] env[63293]: DEBUG nova.compute.manager [req-29b4b8d9-a0ae-461d-967c-75e0473daa7f req-6fb5ee1a-8592-4121-823b-2205dcb27f44 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received event network-vif-plugged-a34b2cc4-afe1-463c-bb22-470ea54d8789 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.420588] env[63293]: DEBUG oslo_concurrency.lockutils [req-29b4b8d9-a0ae-461d-967c-75e0473daa7f req-6fb5ee1a-8592-4121-823b-2205dcb27f44 service nova] Acquiring lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.420794] env[63293]: DEBUG oslo_concurrency.lockutils [req-29b4b8d9-a0ae-461d-967c-75e0473daa7f req-6fb5ee1a-8592-4121-823b-2205dcb27f44 service nova] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.420965] env[63293]: DEBUG oslo_concurrency.lockutils [req-29b4b8d9-a0ae-461d-967c-75e0473daa7f req-6fb5ee1a-8592-4121-823b-2205dcb27f44 service nova] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.421154] env[63293]: DEBUG nova.compute.manager [req-29b4b8d9-a0ae-461d-967c-75e0473daa7f req-6fb5ee1a-8592-4121-823b-2205dcb27f44 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] No waiting events found dispatching network-vif-plugged-a34b2cc4-afe1-463c-bb22-470ea54d8789 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 813.421722] env[63293]: WARNING nova.compute.manager [req-29b4b8d9-a0ae-461d-967c-75e0473daa7f req-6fb5ee1a-8592-4121-823b-2205dcb27f44 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received unexpected event network-vif-plugged-a34b2cc4-afe1-463c-bb22-470ea54d8789 for instance with vm_state building and task_state spawning. [ 813.425948] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327538, 'name': ReconfigVM_Task, 'duration_secs': 0.395835} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.425948] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 0977c79a-2ef3-4ec9-a0a0-de1ed5799723/0977c79a-2ef3-4ec9-a0a0-de1ed5799723.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.426504] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1b8255e-b20a-44ce-bcfa-05a189153f4d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.437434] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 813.437434] env[63293]: value = "task-1327542" [ 813.437434] env[63293]: _type = "Task" [ 813.437434] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.438293] env[63293]: DEBUG nova.network.neutron [req-40d0d556-0635-4234-86f3-7bc67bf74a5e req-1b368108-78e7-4d34-82ac-a4994d85ca68 service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Updated VIF entry in instance network info cache for port e35e7768-ec67-4322-a4c5-be6a5835054e. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 813.438623] env[63293]: DEBUG nova.network.neutron [req-40d0d556-0635-4234-86f3-7bc67bf74a5e req-1b368108-78e7-4d34-82ac-a4994d85ca68 service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Updating instance_info_cache with network_info: [{"id": "e35e7768-ec67-4322-a4c5-be6a5835054e", "address": "fa:16:3e:63:30:1e", "network": {"id": "02c216f2-c41a-4722-91df-330cd520be3d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.39", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f57886847a1446109c6ab061a248dcf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35e7768-ec", "ovs_interfaceid": "e35e7768-ec67-4322-a4c5-be6a5835054e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.441677] env[63293]: DEBUG nova.network.neutron [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Successfully updated port: a34b2cc4-afe1-463c-bb22-470ea54d8789 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 813.453853] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327542, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.662182] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327541, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.666748] env[63293]: DEBUG nova.compute.utils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.671248] env[63293]: DEBUG nova.compute.manager [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 813.671429] env[63293]: DEBUG nova.network.neutron [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 813.750405] env[63293]: DEBUG nova.policy [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6860a9e34a1b44029cf57d9f04ae87c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fac34e49d5dc49e7a7055c998d5b6766', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.878685] env[63293]: DEBUG oslo_vmware.api [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327537, 'name': PowerOnVM_Task, 'duration_secs': 0.546925} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.882644] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 813.882644] env[63293]: INFO nova.compute.manager [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Took 8.48 seconds to spawn the instance on the hypervisor. [ 813.882644] env[63293]: DEBUG nova.compute.manager [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 813.882644] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961ce33b-1737-4373-9186-de18296d9c5d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.931477] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.947666] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.947843] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.949086] env[63293]: DEBUG nova.network.neutron [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.949574] env[63293]: DEBUG oslo_concurrency.lockutils [req-40d0d556-0635-4234-86f3-7bc67bf74a5e req-1b368108-78e7-4d34-82ac-a4994d85ca68 service nova] Releasing lock "refresh_cache-c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.949574] env[63293]: DEBUG nova.compute.manager [req-40d0d556-0635-4234-86f3-7bc67bf74a5e req-1b368108-78e7-4d34-82ac-a4994d85ca68 service nova] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Received event network-vif-deleted-c3bd127f-e525-4481-9814-75afe260e314 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.957471] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327542, 'name': Rename_Task, 'duration_secs': 0.189341} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.957713] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.957946] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15e536fb-2714-4475-866d-a6ac29b8d851 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.969390] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "495cb1b0-9194-44e6-9f68-de62033ea17d" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.969521] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "495cb1b0-9194-44e6-9f68-de62033ea17d" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.969743] env[63293]: INFO nova.compute.manager [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Rebooting instance [ 813.973670] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 813.973670] env[63293]: value = "task-1327543" [ 813.973670] env[63293]: _type = "Task" [ 813.973670] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.990535] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327543, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.039105] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afd5e90-eb60-4f23-b9d3-c8391bb63793 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.046428] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da2feb1-49bf-4c22-a4c1-fce46b331027 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.076733] env[63293]: DEBUG nova.network.neutron [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Successfully created port: 01c58487-3d79-41c6-be70-8e4c5f2efe50 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 814.079124] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7052cd57-9f08-4f19-bc7c-23d1d6fe498a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.086839] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c464645-8cdc-4b98-9f42-3a12b8fbfada {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.100454] env[63293]: DEBUG nova.compute.provider_tree [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.163873] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327541, 'name': CreateVM_Task, 'duration_secs': 0.598461} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.163970] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 814.164653] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.164927] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.165224] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.165471] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfb5697d-50e7-43b7-9893-602ceaa684cf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.170225] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for the task: (returnval){ [ 814.170225] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525c3519-21fa-9fff-9e1a-0bfb89600c4d" [ 814.170225] env[63293]: _type = "Task" [ 814.170225] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.176883] env[63293]: DEBUG nova.compute.manager [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 814.183908] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525c3519-21fa-9fff-9e1a-0bfb89600c4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.400037] env[63293]: INFO nova.compute.manager [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Took 41.99 seconds to build instance. [ 814.488975] env[63293]: DEBUG nova.network.neutron [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.499296] env[63293]: DEBUG nova.compute.manager [req-f8c7a98e-fb03-4f2b-8512-ff7ab8d92507 req-63dc2d22-2b03-463b-8f7f-017f8ca17bf1 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Received event network-changed-70383439-e38d-4f62-8f53-350c7b430342 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.499626] env[63293]: DEBUG nova.compute.manager [req-f8c7a98e-fb03-4f2b-8512-ff7ab8d92507 req-63dc2d22-2b03-463b-8f7f-017f8ca17bf1 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Refreshing instance network info cache due to event network-changed-70383439-e38d-4f62-8f53-350c7b430342. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 814.499909] env[63293]: DEBUG oslo_concurrency.lockutils [req-f8c7a98e-fb03-4f2b-8512-ff7ab8d92507 req-63dc2d22-2b03-463b-8f7f-017f8ca17bf1 service nova] Acquiring lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.504153] env[63293]: DEBUG oslo_concurrency.lockutils [req-f8c7a98e-fb03-4f2b-8512-ff7ab8d92507 req-63dc2d22-2b03-463b-8f7f-017f8ca17bf1 service nova] Acquired lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.504370] env[63293]: DEBUG nova.network.neutron [req-f8c7a98e-fb03-4f2b-8512-ff7ab8d92507 req-63dc2d22-2b03-463b-8f7f-017f8ca17bf1 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Refreshing network info cache for port 70383439-e38d-4f62-8f53-350c7b430342 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 814.510182] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327543, 'name': PowerOnVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.517415] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.606258] env[63293]: DEBUG nova.scheduler.client.report [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.687180] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525c3519-21fa-9fff-9e1a-0bfb89600c4d, 'name': SearchDatastore_Task, 'duration_secs': 0.012626} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.687767] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.688033] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 814.688277] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.688424] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.688599] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.689466] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ecfe78a7-d2f5-40e4-a379-673ec08d9871 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.701854] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.701982] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 814.702737] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b3abc8d-05ce-4088-ac59-511c2378d8ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.706555] env[63293]: DEBUG nova.network.neutron [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [{"id": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "address": "fa:16:3e:2b:1c:64", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b2cc4-af", "ovs_interfaceid": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.715439] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for the task: (returnval){ [ 814.715439] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52bd8ff7-3eb7-dec2-b566-afafecd7de96" [ 814.715439] env[63293]: _type = "Task" [ 814.715439] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.723773] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52bd8ff7-3eb7-dec2-b566-afafecd7de96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.902036] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3dd75c75-e8b1-418d-ae1d-e606816a918c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.028s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.988268] env[63293]: DEBUG oslo_vmware.api [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327543, 'name': PowerOnVM_Task, 'duration_secs': 0.625634} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.988536] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.988738] env[63293]: INFO nova.compute.manager [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Took 7.13 seconds to spawn the instance on the hypervisor. [ 814.988934] env[63293]: DEBUG nova.compute.manager [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.989734] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2358a3d-c5c7-4d6f-9987-521247573d2c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.111157] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.949s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.111745] env[63293]: ERROR nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9276c7bd-1873-447d-aa82-96375b594c8b, please check neutron logs for more information. [ 815.111745] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Traceback (most recent call last): [ 815.111745] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 815.111745] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] self.driver.spawn(context, instance, image_meta, [ 815.111745] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 815.111745] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 815.111745] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 815.111745] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] vm_ref = self.build_virtual_machine(instance, [ 815.111745] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 815.111745] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] vif_infos = vmwarevif.get_vif_info(self._session, [ 815.111745] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] for vif in network_info: [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] return self._sync_wrapper(fn, *args, **kwargs) [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] self.wait() [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] self[:] = self._gt.wait() [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] return self._exit_event.wait() [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] current.throw(*self._exc) [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 815.112127] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] result = function(*args, **kwargs) [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] return func(*args, **kwargs) [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] raise e [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] nwinfo = self.network_api.allocate_for_instance( [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] created_port_ids = self._update_ports_for_instance( [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] with excutils.save_and_reraise_exception(): [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] self.force_reraise() [ 815.112500] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 815.112887] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] raise self.value [ 815.112887] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 815.112887] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] updated_port = self._update_port( [ 815.112887] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 815.112887] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] _ensure_no_port_binding_failure(port) [ 815.112887] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 815.112887] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] raise exception.PortBindingFailed(port_id=port['id']) [ 815.112887] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] nova.exception.PortBindingFailed: Binding failed for port 9276c7bd-1873-447d-aa82-96375b594c8b, please check neutron logs for more information. [ 815.112887] env[63293]: ERROR nova.compute.manager [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] [ 815.112887] env[63293]: DEBUG nova.compute.utils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Binding failed for port 9276c7bd-1873-447d-aa82-96375b594c8b, please check neutron logs for more information. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 815.113933] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.115s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.114899] env[63293]: INFO nova.compute.claims [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.118054] env[63293]: DEBUG nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Build of instance 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4 was re-scheduled: Binding failed for port 9276c7bd-1873-447d-aa82-96375b594c8b, please check neutron logs for more information. {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 815.118484] env[63293]: DEBUG nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Unplugging VIFs for instance {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 815.122014] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Acquiring lock "refresh_cache-52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.122014] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Acquired lock "refresh_cache-52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.122014] env[63293]: DEBUG nova.network.neutron [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.189171] env[63293]: DEBUG nova.compute.manager [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 815.213513] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.213513] env[63293]: DEBUG nova.compute.manager [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Instance network_info: |[{"id": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "address": "fa:16:3e:2b:1c:64", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b2cc4-af", "ovs_interfaceid": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 815.213705] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:1c:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '459b8c74-0aa6-42b6-996a-42b1c5d7e5c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a34b2cc4-afe1-463c-bb22-470ea54d8789', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.221782] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Creating folder: Project (866b347100454019b07f63922b995bb6). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.224473] env[63293]: DEBUG nova.virt.hardware [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 815.224814] env[63293]: DEBUG nova.virt.hardware [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 815.225102] env[63293]: DEBUG nova.virt.hardware [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 815.225405] env[63293]: DEBUG nova.virt.hardware [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 815.225648] env[63293]: DEBUG nova.virt.hardware [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 815.225897] env[63293]: DEBUG nova.virt.hardware [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 815.226705] env[63293]: DEBUG nova.virt.hardware [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 815.226705] env[63293]: DEBUG nova.virt.hardware [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 815.226705] env[63293]: DEBUG nova.virt.hardware [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 815.226705] env[63293]: DEBUG nova.virt.hardware [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 815.226928] env[63293]: DEBUG nova.virt.hardware [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 815.227187] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c827a4df-4f21-490d-8275-a45b0d1bbb3c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.229723] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd9b954-1d51-473b-80ab-313163fa3090 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.245025] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed984779-cef1-426e-8c89-31544035722b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.248776] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52bd8ff7-3eb7-dec2-b566-afafecd7de96, 'name': SearchDatastore_Task, 'duration_secs': 0.009517} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.250818] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Created folder: Project (866b347100454019b07f63922b995bb6) in parent group-v283678. [ 815.251105] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Creating folder: Instances. Parent ref: group-v283717. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.251635] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01397b7d-6a69-4098-9617-b4e2addd00aa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.253547] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aea93396-371e-46ad-989e-7d6cd1568d54 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.265137] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for the task: (returnval){ [ 815.265137] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5215fba7-6104-79f7-06ad-bd06afc52646" [ 815.265137] env[63293]: _type = "Task" [ 815.265137] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.268920] env[63293]: DEBUG nova.network.neutron [req-f8c7a98e-fb03-4f2b-8512-ff7ab8d92507 req-63dc2d22-2b03-463b-8f7f-017f8ca17bf1 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Updated VIF entry in instance network info cache for port 70383439-e38d-4f62-8f53-350c7b430342. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 815.269274] env[63293]: DEBUG nova.network.neutron [req-f8c7a98e-fb03-4f2b-8512-ff7ab8d92507 req-63dc2d22-2b03-463b-8f7f-017f8ca17bf1 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Updating instance_info_cache with network_info: [{"id": "70383439-e38d-4f62-8f53-350c7b430342", "address": "fa:16:3e:ad:36:7b", "network": {"id": "43ed4b72-f94f-47dd-9b14-4297989bd33f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-682241293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3bcac63f7844e10a166eca717f910a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70383439-e3", "ovs_interfaceid": "70383439-e38d-4f62-8f53-350c7b430342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.274226] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5215fba7-6104-79f7-06ad-bd06afc52646, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.274531] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Created folder: Instances in parent group-v283717. [ 815.274757] env[63293]: DEBUG oslo.service.loopingcall [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.275158] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 815.275360] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-07cd80a8-b061-403a-9120-c3b2f0950af9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.295460] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.295460] env[63293]: value = "task-1327546" [ 815.295460] env[63293]: _type = "Task" [ 815.295460] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.302946] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327546, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.404316] env[63293]: DEBUG nova.compute.manager [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 815.469671] env[63293]: DEBUG nova.compute.manager [req-c8477951-0529-4168-ab4b-e931b5151835 req-a4e5d51f-fa8e-45f9-9453-e0e0ef3a531f service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received event network-changed-a34b2cc4-afe1-463c-bb22-470ea54d8789 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 815.469868] env[63293]: DEBUG nova.compute.manager [req-c8477951-0529-4168-ab4b-e931b5151835 req-a4e5d51f-fa8e-45f9-9453-e0e0ef3a531f service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Refreshing instance network info cache due to event network-changed-a34b2cc4-afe1-463c-bb22-470ea54d8789. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 815.470120] env[63293]: DEBUG oslo_concurrency.lockutils [req-c8477951-0529-4168-ab4b-e931b5151835 req-a4e5d51f-fa8e-45f9-9453-e0e0ef3a531f service nova] Acquiring lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.470263] env[63293]: DEBUG oslo_concurrency.lockutils [req-c8477951-0529-4168-ab4b-e931b5151835 req-a4e5d51f-fa8e-45f9-9453-e0e0ef3a531f service nova] Acquired lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.470437] env[63293]: DEBUG nova.network.neutron [req-c8477951-0529-4168-ab4b-e931b5151835 req-a4e5d51f-fa8e-45f9-9453-e0e0ef3a531f service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Refreshing network info cache for port a34b2cc4-afe1-463c-bb22-470ea54d8789 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 815.510318] env[63293]: INFO nova.compute.manager [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Took 41.17 seconds to build instance. [ 815.570819] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65816f58-53cb-4182-a598-301eb7e73375 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.578509] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f344da-ba6a-41f5-8343-a039b6d01748 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Suspending the VM {{(pid=63293) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 815.578846] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-3c1a1db3-23af-45dc-9b0f-30c64ba3ebc1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.585505] env[63293]: DEBUG oslo_vmware.api [None req-a2f344da-ba6a-41f5-8343-a039b6d01748 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 815.585505] env[63293]: value = "task-1327547" [ 815.585505] env[63293]: _type = "Task" [ 815.585505] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.594335] env[63293]: DEBUG oslo_vmware.api [None req-a2f344da-ba6a-41f5-8343-a039b6d01748 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327547, 'name': SuspendVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.644546] env[63293]: DEBUG nova.network.neutron [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.647577] env[63293]: DEBUG nova.network.neutron [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Successfully updated port: 01c58487-3d79-41c6-be70-8e4c5f2efe50 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 815.739173] env[63293]: DEBUG nova.network.neutron [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.775262] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5215fba7-6104-79f7-06ad-bd06afc52646, 'name': SearchDatastore_Task, 'duration_secs': 0.010759} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.775754] env[63293]: DEBUG oslo_concurrency.lockutils [req-f8c7a98e-fb03-4f2b-8512-ff7ab8d92507 req-63dc2d22-2b03-463b-8f7f-017f8ca17bf1 service nova] Releasing lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.776119] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.776409] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] c8009fc0-f7dc-4568-9909-7fa9f7f6a56c/c8009fc0-f7dc-4568-9909-7fa9f7f6a56c.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 815.776635] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquired lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.776843] env[63293]: DEBUG nova.network.neutron [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.778027] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f16ce4ea-467e-4152-9e75-301af54d5691 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.785758] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for the task: (returnval){ [ 815.785758] env[63293]: value = "task-1327548" [ 815.785758] env[63293]: _type = "Task" [ 815.785758] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.795742] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327548, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.804083] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327546, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.928631] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.012551] env[63293]: DEBUG oslo_concurrency.lockutils [None req-aad79462-2e66-4c10-bcf6-265dfbc56d9a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.200s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.097211] env[63293]: DEBUG oslo_vmware.api [None req-a2f344da-ba6a-41f5-8343-a039b6d01748 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327547, 'name': SuspendVM_Task} progress is 58%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.151624] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-33d71260-26f1-482c-b93b-5f8e27c545f3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.151798] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-33d71260-26f1-482c-b93b-5f8e27c545f3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.151964] env[63293]: DEBUG nova.network.neutron [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 816.244190] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Releasing lock "refresh_cache-52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.244190] env[63293]: DEBUG nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63293) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 816.244190] env[63293]: DEBUG nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 816.244190] env[63293]: DEBUG nova.network.neutron [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.297615] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327548, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496077} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.301479] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] c8009fc0-f7dc-4568-9909-7fa9f7f6a56c/c8009fc0-f7dc-4568-9909-7fa9f7f6a56c.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 816.301736] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 816.301938] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c371f6b-b7ec-44a0-8b44-af2a80c19587 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.309763] env[63293]: DEBUG nova.network.neutron [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.319241] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for the task: (returnval){ [ 816.319241] env[63293]: value = "task-1327549" [ 816.319241] env[63293]: _type = "Task" [ 816.319241] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.319450] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327546, 'name': CreateVM_Task, 'duration_secs': 0.523225} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.324688] env[63293]: DEBUG nova.network.neutron [req-c8477951-0529-4168-ab4b-e931b5151835 req-a4e5d51f-fa8e-45f9-9453-e0e0ef3a531f service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updated VIF entry in instance network info cache for port a34b2cc4-afe1-463c-bb22-470ea54d8789. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 816.325464] env[63293]: DEBUG nova.network.neutron [req-c8477951-0529-4168-ab4b-e931b5151835 req-a4e5d51f-fa8e-45f9-9453-e0e0ef3a531f service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [{"id": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "address": "fa:16:3e:2b:1c:64", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b2cc4-af", "ovs_interfaceid": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.327938] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 816.332974] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.333155] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.333863] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.334447] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76810643-2d53-4259-93d4-92fce56becf4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.342554] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327549, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.347576] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 816.347576] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5231112a-b133-06cf-06c2-425d69b3e18b" [ 816.347576] env[63293]: _type = "Task" [ 816.347576] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.357558] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5231112a-b133-06cf-06c2-425d69b3e18b, 'name': SearchDatastore_Task, 'duration_secs': 0.010118} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.357750] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.357910] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.358147] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.358290] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.358464] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.361125] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db75ba0b-4940-4ccf-a9d4-6a8d2c632ceb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.369134] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.369319] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 816.370159] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30035b28-0229-4731-81b9-71d61a39b07a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.376723] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 816.376723] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52be347d-9049-a5b1-5631-a6cb887496b8" [ 816.376723] env[63293]: _type = "Task" [ 816.376723] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.388485] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52be347d-9049-a5b1-5631-a6cb887496b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.487615] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7daf8478-4206-4e85-93ae-d51a96127c01 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.497073] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247dc68d-11ba-414f-8838-017d047375fc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.526539] env[63293]: DEBUG nova.compute.manager [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 816.532068] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1f5c92-1e5b-41f6-8388-38fd1c54c9bd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.540015] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd622836-a608-4dff-b90d-cfd48ca041ae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.554440] env[63293]: DEBUG nova.compute.provider_tree [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.597420] env[63293]: DEBUG oslo_vmware.api [None req-a2f344da-ba6a-41f5-8343-a039b6d01748 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327547, 'name': SuspendVM_Task, 'duration_secs': 0.716838} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.597685] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f344da-ba6a-41f5-8343-a039b6d01748 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Suspended the VM {{(pid=63293) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 816.597873] env[63293]: DEBUG nova.compute.manager [None req-a2f344da-ba6a-41f5-8343-a039b6d01748 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.598696] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620ec74d-10ac-423b-a2eb-eadd4a56eb64 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.697899] env[63293]: DEBUG nova.network.neutron [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.791919] env[63293]: DEBUG nova.network.neutron [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Updating instance_info_cache with network_info: [{"id": "70383439-e38d-4f62-8f53-350c7b430342", "address": "fa:16:3e:ad:36:7b", "network": {"id": "43ed4b72-f94f-47dd-9b14-4297989bd33f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-682241293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3bcac63f7844e10a166eca717f910a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70383439-e3", "ovs_interfaceid": "70383439-e38d-4f62-8f53-350c7b430342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.828469] env[63293]: DEBUG nova.network.neutron [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.830043] env[63293]: DEBUG oslo_concurrency.lockutils [req-c8477951-0529-4168-ab4b-e931b5151835 req-a4e5d51f-fa8e-45f9-9453-e0e0ef3a531f service nova] Releasing lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.835056] env[63293]: DEBUG nova.network.neutron [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Updating instance_info_cache with network_info: [{"id": "01c58487-3d79-41c6-be70-8e4c5f2efe50", "address": "fa:16:3e:ec:ac:e5", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01c58487-3d", "ovs_interfaceid": "01c58487-3d79-41c6-be70-8e4c5f2efe50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.839445] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327549, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108521} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.840234] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 816.841082] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84c1ecf-f9a3-4b91-bdb8-3c8a328c677a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.865697] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] c8009fc0-f7dc-4568-9909-7fa9f7f6a56c/c8009fc0-f7dc-4568-9909-7fa9f7f6a56c.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.866335] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a88956ec-b0e0-45a7-902e-8efd855af75b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.897117] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52be347d-9049-a5b1-5631-a6cb887496b8, 'name': SearchDatastore_Task, 'duration_secs': 0.018852} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.899085] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for the task: (returnval){ [ 816.899085] env[63293]: value = "task-1327550" [ 816.899085] env[63293]: _type = "Task" [ 816.899085] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.899269] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-956d64dd-2e04-4616-9d9d-c95e51cf075f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.907143] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 816.907143] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e096c5-0360-93f4-27bf-11921e39415f" [ 816.907143] env[63293]: _type = "Task" [ 816.907143] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.910414] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327550, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.917505] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e096c5-0360-93f4-27bf-11921e39415f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.057213] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.058101] env[63293]: DEBUG nova.scheduler.client.report [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.296665] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Releasing lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.298538] env[63293]: DEBUG nova.compute.manager [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.298663] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674bdab9-caf8-4308-9bba-128ee4dbdb1d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.335571] env[63293]: INFO nova.compute.manager [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] [instance: 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4] Took 1.09 seconds to deallocate network for instance. [ 817.339568] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-33d71260-26f1-482c-b93b-5f8e27c545f3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.339568] env[63293]: DEBUG nova.compute.manager [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Instance network_info: |[{"id": "01c58487-3d79-41c6-be70-8e4c5f2efe50", "address": "fa:16:3e:ec:ac:e5", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01c58487-3d", "ovs_interfaceid": "01c58487-3d79-41c6-be70-8e4c5f2efe50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 817.339928] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:ac:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aec0089a-ff85-4bef-bad8-c84de39af71a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01c58487-3d79-41c6-be70-8e4c5f2efe50', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 817.348533] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Creating folder: Project (fac34e49d5dc49e7a7055c998d5b6766). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 817.349625] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7bc35848-a0c6-446e-bb85-ee73c77f7ff5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.360137] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Created folder: Project (fac34e49d5dc49e7a7055c998d5b6766) in parent group-v283678. [ 817.360278] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Creating folder: Instances. Parent ref: group-v283720. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 817.360515] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-61955b5d-d9ef-406b-bc9d-a9e5351b5b9f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.369611] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Created folder: Instances in parent group-v283720. [ 817.369858] env[63293]: DEBUG oslo.service.loopingcall [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.370137] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 817.370387] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-373d69b9-ec37-4db3-9720-24552bd8a4b2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.390874] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 817.390874] env[63293]: value = "task-1327553" [ 817.390874] env[63293]: _type = "Task" [ 817.390874] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.399296] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327553, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.409646] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327550, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.420816] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e096c5-0360-93f4-27bf-11921e39415f, 'name': SearchDatastore_Task, 'duration_secs': 0.010124} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.421456] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.421717] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 65c2f72d-6bbe-45ac-8efc-401e0664390f/65c2f72d-6bbe-45ac-8efc-401e0664390f.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 817.422031] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e364d60a-620b-43ff-b824-4fcdb2abd8da {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.428540] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 817.428540] env[63293]: value = "task-1327554" [ 817.428540] env[63293]: _type = "Task" [ 817.428540] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.436475] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327554, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.548053] env[63293]: DEBUG nova.compute.manager [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Received event network-vif-plugged-01c58487-3d79-41c6-be70-8e4c5f2efe50 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.548306] env[63293]: DEBUG oslo_concurrency.lockutils [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] Acquiring lock "33d71260-26f1-482c-b93b-5f8e27c545f3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.548474] env[63293]: DEBUG oslo_concurrency.lockutils [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] Lock "33d71260-26f1-482c-b93b-5f8e27c545f3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.548755] env[63293]: DEBUG oslo_concurrency.lockutils [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] Lock "33d71260-26f1-482c-b93b-5f8e27c545f3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.548755] env[63293]: DEBUG nova.compute.manager [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] No waiting events found dispatching network-vif-plugged-01c58487-3d79-41c6-be70-8e4c5f2efe50 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 817.548973] env[63293]: WARNING nova.compute.manager [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Received unexpected event network-vif-plugged-01c58487-3d79-41c6-be70-8e4c5f2efe50 for instance with vm_state building and task_state spawning. [ 817.549198] env[63293]: DEBUG nova.compute.manager [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Received event network-changed-01c58487-3d79-41c6-be70-8e4c5f2efe50 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.549400] env[63293]: DEBUG nova.compute.manager [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Refreshing instance network info cache due to event network-changed-01c58487-3d79-41c6-be70-8e4c5f2efe50. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.549621] env[63293]: DEBUG oslo_concurrency.lockutils [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] Acquiring lock "refresh_cache-33d71260-26f1-482c-b93b-5f8e27c545f3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.549759] env[63293]: DEBUG oslo_concurrency.lockutils [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] Acquired lock "refresh_cache-33d71260-26f1-482c-b93b-5f8e27c545f3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.549917] env[63293]: DEBUG nova.network.neutron [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Refreshing network info cache for port 01c58487-3d79-41c6-be70-8e4c5f2efe50 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 817.566496] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.566875] env[63293]: DEBUG nova.compute.manager [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 817.569779] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.672s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.571453] env[63293]: INFO nova.compute.claims [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.901482] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327553, 'name': CreateVM_Task, 'duration_secs': 0.429042} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.901768] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 817.905471] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.905669] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.906012] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.906283] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85693d34-5dc5-4f02-be52-ff5dcd1aefdd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.914542] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327550, 'name': ReconfigVM_Task, 'duration_secs': 0.605765} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.916169] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Reconfigured VM instance instance-0000003a to attach disk [datastore1] c8009fc0-f7dc-4568-9909-7fa9f7f6a56c/c8009fc0-f7dc-4568-9909-7fa9f7f6a56c.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 817.916769] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 817.916769] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e234cb-8a39-c69a-0f48-5123e4238573" [ 817.916769] env[63293]: _type = "Task" [ 817.916769] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.916977] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e9ca74d-c0d7-4c40-b29e-adcc0e61a24b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.926983] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e234cb-8a39-c69a-0f48-5123e4238573, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.929473] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for the task: (returnval){ [ 817.929473] env[63293]: value = "task-1327555" [ 817.929473] env[63293]: _type = "Task" [ 817.929473] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.941845] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327554, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.944977] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327555, 'name': Rename_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.078674] env[63293]: DEBUG nova.compute.utils [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.083584] env[63293]: DEBUG nova.compute.manager [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 818.083584] env[63293]: DEBUG nova.network.neutron [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 818.135115] env[63293]: DEBUG nova.policy [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf7a8a464fd54245b66ed9cd75743779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7fd63e6892c44f159195897ad5eacf13', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 818.314957] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d4a917-125f-4aaa-9d00-1663456e99c6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.323224] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Doing hard reboot of VM {{(pid=63293) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 818.323490] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-fb37521e-2e65-4513-b7ed-a714e975c86e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.330165] env[63293]: DEBUG oslo_vmware.api [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for the task: (returnval){ [ 818.330165] env[63293]: value = "task-1327556" [ 818.330165] env[63293]: _type = "Task" [ 818.330165] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.339267] env[63293]: DEBUG oslo_vmware.api [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327556, 'name': ResetVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.341237] env[63293]: DEBUG nova.network.neutron [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Updated VIF entry in instance network info cache for port 01c58487-3d79-41c6-be70-8e4c5f2efe50. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 818.341737] env[63293]: DEBUG nova.network.neutron [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Updating instance_info_cache with network_info: [{"id": "01c58487-3d79-41c6-be70-8e4c5f2efe50", "address": "fa:16:3e:ec:ac:e5", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01c58487-3d", "ovs_interfaceid": "01c58487-3d79-41c6-be70-8e4c5f2efe50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.379114] env[63293]: INFO nova.scheduler.client.report [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Deleted allocations for instance 52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4 [ 818.431016] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e234cb-8a39-c69a-0f48-5123e4238573, 'name': SearchDatastore_Task, 'duration_secs': 0.028479} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.434517] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.435289] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 818.435289] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.435289] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.435588] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 818.436200] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9d3e9062-29c0-4e7a-a4b7-ec66affbfa7a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.447043] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327554, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.450561] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 65c2f72d-6bbe-45ac-8efc-401e0664390f/65c2f72d-6bbe-45ac-8efc-401e0664390f.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 818.450790] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.451717] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327555, 'name': Rename_Task, 'duration_secs': 0.45943} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.452346] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0426e286-461c-4e73-8f2f-32bed01772b7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.454404] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 818.454676] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 818.454837] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 818.455649] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ff8050e-59b4-4d2f-9c6f-c70b705dd09a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.458151] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04987dc7-882f-4368-a207-c7156d18cd82 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.464628] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 818.464628] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5250b890-8531-39aa-6927-2ccfb7a22a47" [ 818.464628] env[63293]: _type = "Task" [ 818.464628] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.467120] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for the task: (returnval){ [ 818.467120] env[63293]: value = "task-1327558" [ 818.467120] env[63293]: _type = "Task" [ 818.467120] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.467364] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 818.467364] env[63293]: value = "task-1327557" [ 818.467364] env[63293]: _type = "Task" [ 818.467364] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.485192] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5250b890-8531-39aa-6927-2ccfb7a22a47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.490733] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327558, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.491111] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327557, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.542202] env[63293]: DEBUG nova.network.neutron [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Successfully created port: aa552093-dd1a-4dd1-aadf-9c92b766b4ea {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.585681] env[63293]: DEBUG nova.compute.manager [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 818.847360] env[63293]: DEBUG oslo_concurrency.lockutils [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] Releasing lock "refresh_cache-33d71260-26f1-482c-b93b-5f8e27c545f3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.849016] env[63293]: DEBUG nova.compute.manager [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Received event network-changed-2f0bfb1c-6184-49f1-9c71-cb7459e1470d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.849016] env[63293]: DEBUG nova.compute.manager [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Refreshing instance network info cache due to event network-changed-2f0bfb1c-6184-49f1-9c71-cb7459e1470d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 818.849016] env[63293]: DEBUG oslo_concurrency.lockutils [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] Acquiring lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.849016] env[63293]: DEBUG oslo_concurrency.lockutils [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] Acquired lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.849016] env[63293]: DEBUG nova.network.neutron [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Refreshing network info cache for port 2f0bfb1c-6184-49f1-9c71-cb7459e1470d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.849401] env[63293]: DEBUG oslo_vmware.api [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327556, 'name': ResetVM_Task, 'duration_secs': 0.107175} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.855638] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Did hard reboot of VM {{(pid=63293) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 818.855827] env[63293]: DEBUG nova.compute.manager [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.860727] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296927c4-1fc6-4c0c-b6d3-adfaa63e55a0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.890145] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1145536-2e19-48b0-b83e-e50d1b477987 tempest-InstanceActionsNegativeTestJSON-728538985 tempest-InstanceActionsNegativeTestJSON-728538985-project-member] Lock "52fe2bd1-7b43-4c11-b2b4-3aa35d508ae4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.380s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.928984] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defd4889-c106-44e0-bbfe-2b664f05e767 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.938823] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d653a00-ffb6-4275-99d6-556aa001e68a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.982403] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b71ec2b5-b459-4061-b792-42a9867a3877 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.990718] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5250b890-8531-39aa-6927-2ccfb7a22a47, 'name': SearchDatastore_Task, 'duration_secs': 0.026515} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.999564] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327557, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.146954} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.999867] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327558, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.000119] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4480201c-f444-4d24-a1c7-c012076769f3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.003864] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62661cc-839f-4d3f-87fc-ab22fd4b876f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.007821] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.008889] env[63293]: DEBUG nova.compute.manager [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.009598] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc1b670-66dd-408c-91be-1083bdfad10b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.012308] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d78830-7192-4e4e-b4b1-f677d6222d85 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.017853] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 819.017853] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]526b59e0-75fc-3528-cdd2-a504e223f672" [ 819.017853] env[63293]: _type = "Task" [ 819.017853] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.044880] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 65c2f72d-6bbe-45ac-8efc-401e0664390f/65c2f72d-6bbe-45ac-8efc-401e0664390f.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.047707] env[63293]: DEBUG nova.compute.provider_tree [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.053525] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da05b381-4beb-4a31-adce-f69c93de5c5f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.069856] env[63293]: DEBUG nova.scheduler.client.report [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.079413] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]526b59e0-75fc-3528-cdd2-a504e223f672, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.080925] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 819.080925] env[63293]: value = "task-1327559" [ 819.080925] env[63293]: _type = "Task" [ 819.080925] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.090133] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327559, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.092850] env[63293]: INFO nova.virt.block_device [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Booting with volume 1f5a039e-f604-47d3-a580-f1754a5b697d at /dev/sda [ 819.138423] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e61d020-319a-4ac3-9f2c-56c19e764a93 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.151398] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60837a00-5ab7-4aa3-9bc5-52d5e4662291 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.179570] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c44652e-41a0-49e4-8ab7-7ce96116947e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.186021] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc49d9b-09fb-4bee-8cae-2f94b3589ba2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.212766] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de8d603-7796-43f6-9a95-62b27629fcc1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.219510] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56df57df-5a6b-4bbc-b659-93657373a333 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.233054] env[63293]: DEBUG nova.virt.block_device [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Updating existing volume attachment record: bbfde594-d781-4afc-8141-496cc4f49ab2 {{(pid=63293) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 819.382081] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8231d7cd-9d71-4697-9222-9f8f30c7c258 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "495cb1b0-9194-44e6-9f68-de62033ea17d" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.412s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.393219] env[63293]: DEBUG nova.compute.manager [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 819.485218] env[63293]: DEBUG oslo_vmware.api [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327558, 'name': PowerOnVM_Task, 'duration_secs': 0.955159} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.485502] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 819.485706] env[63293]: INFO nova.compute.manager [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Took 9.23 seconds to spawn the instance on the hypervisor. [ 819.486147] env[63293]: DEBUG nova.compute.manager [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.486737] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36753e23-3fb8-48db-b497-c100ca8c9603 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.547729] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]526b59e0-75fc-3528-cdd2-a504e223f672, 'name': SearchDatastore_Task, 'duration_secs': 0.062981} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.547729] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.547963] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 33d71260-26f1-482c-b93b-5f8e27c545f3/33d71260-26f1-482c-b93b-5f8e27c545f3.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 819.548311] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a59af20f-349c-4bc4-8a95-1ead7935013c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.552093] env[63293]: INFO nova.compute.manager [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] instance snapshotting [ 819.552173] env[63293]: WARNING nova.compute.manager [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 819.558756] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf71292a-32f2-4757-a6f2-54add8cc3078 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.563524] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 819.563524] env[63293]: value = "task-1327560" [ 819.563524] env[63293]: _type = "Task" [ 819.563524] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.587142] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.017s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.587817] env[63293]: DEBUG nova.compute.manager [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.591458] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.910s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.593750] env[63293]: INFO nova.compute.claims [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.600036] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638fa34c-593f-4c3f-a1df-5db585a9d689 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.607692] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327560, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.614912] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327559, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.760677] env[63293]: DEBUG nova.network.neutron [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updated VIF entry in instance network info cache for port 2f0bfb1c-6184-49f1-9c71-cb7459e1470d. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 819.761073] env[63293]: DEBUG nova.network.neutron [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updating instance_info_cache with network_info: [{"id": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "address": "fa:16:3e:80:11:9c", "network": {"id": "9d5a3e58-567a-4c3f-9ad9-9fcbf81fced9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-949261114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1a30d37c1b47fd8670db0a9ef7020d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f0bfb1c-61", "ovs_interfaceid": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.923796] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.010145] env[63293]: INFO nova.compute.manager [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Took 43.70 seconds to build instance. [ 820.074158] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327560, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.101888] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327559, 'name': ReconfigVM_Task, 'duration_secs': 0.5917} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.102204] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 65c2f72d-6bbe-45ac-8efc-401e0664390f/65c2f72d-6bbe-45ac-8efc-401e0664390f.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.102843] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c56f096b-4e6e-4d1c-88e8-c9b6610eb110 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.109371] env[63293]: DEBUG nova.compute.utils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.110930] env[63293]: DEBUG nova.compute.manager [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.110930] env[63293]: DEBUG nova.network.neutron [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 820.114239] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 820.114239] env[63293]: value = "task-1327561" [ 820.114239] env[63293]: _type = "Task" [ 820.114239] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.121924] env[63293]: DEBUG nova.compute.manager [req-9183fc9e-0de3-40b6-a0ae-82367cc45a99 req-e951c196-653b-48ac-ab31-6a794838cdfb service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Received event network-changed-70383439-e38d-4f62-8f53-350c7b430342 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.122138] env[63293]: DEBUG nova.compute.manager [req-9183fc9e-0de3-40b6-a0ae-82367cc45a99 req-e951c196-653b-48ac-ab31-6a794838cdfb service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Refreshing instance network info cache due to event network-changed-70383439-e38d-4f62-8f53-350c7b430342. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 820.122436] env[63293]: DEBUG oslo_concurrency.lockutils [req-9183fc9e-0de3-40b6-a0ae-82367cc45a99 req-e951c196-653b-48ac-ab31-6a794838cdfb service nova] Acquiring lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.122525] env[63293]: DEBUG oslo_concurrency.lockutils [req-9183fc9e-0de3-40b6-a0ae-82367cc45a99 req-e951c196-653b-48ac-ab31-6a794838cdfb service nova] Acquired lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.122635] env[63293]: DEBUG nova.network.neutron [req-9183fc9e-0de3-40b6-a0ae-82367cc45a99 req-e951c196-653b-48ac-ab31-6a794838cdfb service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Refreshing network info cache for port 70383439-e38d-4f62-8f53-350c7b430342 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 820.127556] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Creating Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 820.134690] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b7290dc8-ff3e-4291-864b-356d50153139 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.138660] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327561, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.144021] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 820.144021] env[63293]: value = "task-1327562" [ 820.144021] env[63293]: _type = "Task" [ 820.144021] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.157751] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327562, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.226593] env[63293]: DEBUG nova.policy [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32641f4d76e24c6b809af9f005b2dab7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6fa6cd8dea16412db105faf2ff93a6ac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 820.264427] env[63293]: DEBUG oslo_concurrency.lockutils [req-f1acf82f-17c0-420d-b969-b2c23ed24601 req-6e2a5039-5fdc-4f24-9d88-695bf7311ab1 service nova] Releasing lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.384402] env[63293]: DEBUG nova.network.neutron [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Successfully updated port: aa552093-dd1a-4dd1-aadf-9c92b766b4ea {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.513037] env[63293]: DEBUG oslo_concurrency.lockutils [None req-87298d8a-9d0f-46fa-9ea4-4cb2e8bd6e59 tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.869s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.575989] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327560, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.817091} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.576778] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 33d71260-26f1-482c-b93b-5f8e27c545f3/33d71260-26f1-482c-b93b-5f8e27c545f3.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 820.579020] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.579020] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d76b6c81-eb1b-4b0f-a1a3-20cc78dc4294 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.588703] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 820.588703] env[63293]: value = "task-1327563" [ 820.588703] env[63293]: _type = "Task" [ 820.588703] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.598810] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327563, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.616032] env[63293]: DEBUG nova.compute.manager [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.636759] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327561, 'name': Rename_Task, 'duration_secs': 0.165085} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.639455] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 820.640159] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c58f9a3-fb51-4de9-a9d5-30c3de09aa6c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.649871] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 820.649871] env[63293]: value = "task-1327564" [ 820.649871] env[63293]: _type = "Task" [ 820.649871] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.660111] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327562, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.667089] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327564, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.889583] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Acquiring lock "refresh_cache-b8637881-9fb6-4c5d-848d-7b2d38f8e970" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.889731] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Acquired lock "refresh_cache-b8637881-9fb6-4c5d-848d-7b2d38f8e970" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.889881] env[63293]: DEBUG nova.network.neutron [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.922304] env[63293]: DEBUG nova.network.neutron [req-9183fc9e-0de3-40b6-a0ae-82367cc45a99 req-e951c196-653b-48ac-ab31-6a794838cdfb service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Updated VIF entry in instance network info cache for port 70383439-e38d-4f62-8f53-350c7b430342. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 820.926562] env[63293]: DEBUG nova.network.neutron [req-9183fc9e-0de3-40b6-a0ae-82367cc45a99 req-e951c196-653b-48ac-ab31-6a794838cdfb service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Updating instance_info_cache with network_info: [{"id": "70383439-e38d-4f62-8f53-350c7b430342", "address": "fa:16:3e:ad:36:7b", "network": {"id": "43ed4b72-f94f-47dd-9b14-4297989bd33f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-682241293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3bcac63f7844e10a166eca717f910a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70383439-e3", "ovs_interfaceid": "70383439-e38d-4f62-8f53-350c7b430342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.941985] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81412bc-bdb7-452d-9a7b-2873dbcc5a31 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.951597] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fb0e2d-179b-4f67-870e-07f86b1e9c7c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.955678] env[63293]: DEBUG nova.compute.manager [None req-97827eb8-1508-4cd4-aebe-2599a1a66bde tempest-ServerDiagnosticsTest-1458113536 tempest-ServerDiagnosticsTest-1458113536-project-admin] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.957053] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5a283d-a2c6-4b1a-a327-9d0b703cf433 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.963408] env[63293]: INFO nova.compute.manager [None req-97827eb8-1508-4cd4-aebe-2599a1a66bde tempest-ServerDiagnosticsTest-1458113536 tempest-ServerDiagnosticsTest-1458113536-project-admin] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Retrieving diagnostics [ 820.988606] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14da8744-ae0c-4f14-a8d7-f8b2c14da010 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.992872] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c55251-e2f7-456a-8018-4b53e5291a7b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.995418] env[63293]: DEBUG oslo_concurrency.lockutils [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "495cb1b0-9194-44e6-9f68-de62033ea17d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.995647] env[63293]: DEBUG oslo_concurrency.lockutils [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "495cb1b0-9194-44e6-9f68-de62033ea17d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.995846] env[63293]: DEBUG oslo_concurrency.lockutils [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "495cb1b0-9194-44e6-9f68-de62033ea17d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.996016] env[63293]: DEBUG oslo_concurrency.lockutils [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "495cb1b0-9194-44e6-9f68-de62033ea17d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.996183] env[63293]: DEBUG oslo_concurrency.lockutils [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "495cb1b0-9194-44e6-9f68-de62033ea17d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.998337] env[63293]: DEBUG nova.network.neutron [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Successfully created port: 26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.000927] env[63293]: INFO nova.compute.manager [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Terminating instance [ 821.026728] env[63293]: DEBUG nova.compute.manager [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 821.027033] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 821.027784] env[63293]: DEBUG nova.compute.manager [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.032046] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c67ff18-eb21-403d-b9bd-f332b3ddd673 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.038931] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ec7469-2b22-4961-957d-44aef468e58c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.045377] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 821.045975] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2df7ab64-6f30-4029-ab00-35c7aa2d19f8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.055110] env[63293]: DEBUG nova.compute.provider_tree [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.061215] env[63293]: DEBUG oslo_vmware.api [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for the task: (returnval){ [ 821.061215] env[63293]: value = "task-1327565" [ 821.061215] env[63293]: _type = "Task" [ 821.061215] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.070052] env[63293]: DEBUG oslo_vmware.api [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327565, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.096568] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327563, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110639} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.096820] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.097641] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dec6add-3039-402c-b701-f0e959d7f836 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.121577] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 33d71260-26f1-482c-b93b-5f8e27c545f3/33d71260-26f1-482c-b93b-5f8e27c545f3.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.121935] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-463f2773-6fbe-4be2-80f4-1e30df6ff6d7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.146497] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 821.146497] env[63293]: value = "task-1327566" [ 821.146497] env[63293]: _type = "Task" [ 821.146497] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.158623] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327562, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.162071] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327566, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.166047] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327564, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.358997] env[63293]: DEBUG nova.compute.manager [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.359644] env[63293]: DEBUG nova.virt.hardware [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.359885] env[63293]: DEBUG nova.virt.hardware [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.360054] env[63293]: DEBUG nova.virt.hardware [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.360246] env[63293]: DEBUG nova.virt.hardware [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.360390] env[63293]: DEBUG nova.virt.hardware [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.360535] env[63293]: DEBUG nova.virt.hardware [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.360786] env[63293]: DEBUG nova.virt.hardware [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.361031] env[63293]: DEBUG nova.virt.hardware [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.361272] env[63293]: DEBUG nova.virt.hardware [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.361522] env[63293]: DEBUG nova.virt.hardware [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.361774] env[63293]: DEBUG nova.virt.hardware [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.362678] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-095bfedc-d863-49f5-99e9-34b48628da49 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.371973] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e8ceb1-7278-47c6-89e3-d8fa87ab33ea {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.421428] env[63293]: DEBUG nova.network.neutron [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.428325] env[63293]: DEBUG oslo_concurrency.lockutils [req-9183fc9e-0de3-40b6-a0ae-82367cc45a99 req-e951c196-653b-48ac-ab31-6a794838cdfb service nova] Releasing lock "refresh_cache-495cb1b0-9194-44e6-9f68-de62033ea17d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.555621] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.557901] env[63293]: DEBUG nova.scheduler.client.report [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.582563] env[63293]: DEBUG oslo_vmware.api [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327565, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.643182] env[63293]: DEBUG nova.compute.manager [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.651673] env[63293]: DEBUG nova.network.neutron [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Updating instance_info_cache with network_info: [{"id": "aa552093-dd1a-4dd1-aadf-9c92b766b4ea", "address": "fa:16:3e:13:f8:5d", "network": {"id": "0ec4e867-7095-4c29-9202-669ad2701e74", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-378057015-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fd63e6892c44f159195897ad5eacf13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa552093-dd", "ovs_interfaceid": "aa552093-dd1a-4dd1-aadf-9c92b766b4ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.663663] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327562, 'name': CreateSnapshot_Task, 'duration_secs': 1.484634} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.664258] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Created Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 821.665032] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b2080c-ea19-4e1f-93f2-59d4416eba74 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.674230] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327564, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.674459] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327566, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.688982] env[63293]: DEBUG nova.virt.hardware [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.689303] env[63293]: DEBUG nova.virt.hardware [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.689505] env[63293]: DEBUG nova.virt.hardware [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.689723] env[63293]: DEBUG nova.virt.hardware [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.689896] env[63293]: DEBUG nova.virt.hardware [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.690096] env[63293]: DEBUG nova.virt.hardware [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.690313] env[63293]: DEBUG nova.virt.hardware [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.690473] env[63293]: DEBUG nova.virt.hardware [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.690642] env[63293]: DEBUG nova.virt.hardware [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.690802] env[63293]: DEBUG nova.virt.hardware [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.690976] env[63293]: DEBUG nova.virt.hardware [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.692175] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8846a7-9e8f-43d6-b042-88ab29809efa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.701703] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912db54a-b458-436f-a860-3366744a599b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.070555] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.071107] env[63293]: DEBUG nova.compute.manager [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 822.078452] env[63293]: DEBUG oslo_concurrency.lockutils [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.770s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.078755] env[63293]: DEBUG nova.objects.instance [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lazy-loading 'resources' on Instance uuid b48324a4-902d-4642-8827-daee9683b3a0 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 822.087718] env[63293]: DEBUG oslo_vmware.api [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327565, 'name': PowerOffVM_Task, 'duration_secs': 1.000766} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.087928] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 822.088233] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 822.091133] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ddee701-b1dc-4b15-8c9b-315f86be2357 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.149682] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 822.149954] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 822.150169] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Deleting the datastore file [datastore2] 495cb1b0-9194-44e6-9f68-de62033ea17d {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 822.150436] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15ffd742-d997-4c5a-abbf-57bcd46000ab {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.162733] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Releasing lock "refresh_cache-b8637881-9fb6-4c5d-848d-7b2d38f8e970" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.163106] env[63293]: DEBUG nova.compute.manager [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Instance network_info: |[{"id": "aa552093-dd1a-4dd1-aadf-9c92b766b4ea", "address": "fa:16:3e:13:f8:5d", "network": {"id": "0ec4e867-7095-4c29-9202-669ad2701e74", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-378057015-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fd63e6892c44f159195897ad5eacf13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa552093-dd", "ovs_interfaceid": "aa552093-dd1a-4dd1-aadf-9c92b766b4ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 822.163580] env[63293]: DEBUG oslo_vmware.api [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for the task: (returnval){ [ 822.163580] env[63293]: value = "task-1327568" [ 822.163580] env[63293]: _type = "Task" [ 822.163580] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.163909] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:f8:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3a80436-f7a9-431a-acec-aca3d76e3f9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa552093-dd1a-4dd1-aadf-9c92b766b4ea', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.172583] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Creating folder: Project (7fd63e6892c44f159195897ad5eacf13). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.176028] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d770ea2f-5c28-4420-a8d4-fb0e68e1de55 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.184435] env[63293]: DEBUG oslo_vmware.api [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327564, 'name': PowerOnVM_Task, 'duration_secs': 1.191543} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.192285] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Creating linked-clone VM from snapshot {{(pid=63293) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 822.192636] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327566, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.193712] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 822.193941] env[63293]: INFO nova.compute.manager [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Took 9.48 seconds to spawn the instance on the hypervisor. [ 822.194227] env[63293]: DEBUG nova.compute.manager [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.198021] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-061f214c-cf05-4c2a-a77f-af6003e8e75d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.202355] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d13245-f2cb-47d6-9c6d-8854692c0aee {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.205541] env[63293]: DEBUG oslo_vmware.api [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327568, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.207465] env[63293]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 822.207690] env[63293]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63293) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 822.209625] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Folder already exists: Project (7fd63e6892c44f159195897ad5eacf13). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 822.209833] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Creating folder: Instances. Parent ref: group-v283691. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.211461] env[63293]: DEBUG nova.compute.manager [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Received event network-vif-plugged-aa552093-dd1a-4dd1-aadf-9c92b766b4ea {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.211608] env[63293]: DEBUG oslo_concurrency.lockutils [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] Acquiring lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.211828] env[63293]: DEBUG oslo_concurrency.lockutils [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] Lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.212025] env[63293]: DEBUG oslo_concurrency.lockutils [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] Lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.212207] env[63293]: DEBUG nova.compute.manager [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] No waiting events found dispatching network-vif-plugged-aa552093-dd1a-4dd1-aadf-9c92b766b4ea {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 822.212367] env[63293]: WARNING nova.compute.manager [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Received unexpected event network-vif-plugged-aa552093-dd1a-4dd1-aadf-9c92b766b4ea for instance with vm_state building and task_state spawning. [ 822.212537] env[63293]: DEBUG nova.compute.manager [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Received event network-changed-aa552093-dd1a-4dd1-aadf-9c92b766b4ea {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.212696] env[63293]: DEBUG nova.compute.manager [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Refreshing instance network info cache due to event network-changed-aa552093-dd1a-4dd1-aadf-9c92b766b4ea. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.213166] env[63293]: DEBUG oslo_concurrency.lockutils [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] Acquiring lock "refresh_cache-b8637881-9fb6-4c5d-848d-7b2d38f8e970" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.213320] env[63293]: DEBUG oslo_concurrency.lockutils [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] Acquired lock "refresh_cache-b8637881-9fb6-4c5d-848d-7b2d38f8e970" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.213477] env[63293]: DEBUG nova.network.neutron [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Refreshing network info cache for port aa552093-dd1a-4dd1-aadf-9c92b766b4ea {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.216709] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a174c3f8-b845-4f57-9f46-b8f52227d736 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.218822] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 822.218822] env[63293]: value = "task-1327570" [ 822.218822] env[63293]: _type = "Task" [ 822.218822] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.234829] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327570, 'name': CloneVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.236463] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Created folder: Instances in parent group-v283691. [ 822.236702] env[63293]: DEBUG oslo.service.loopingcall [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.236880] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 822.237401] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-90b78b16-f200-4a63-8cd8-c13ee7bdc224 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.256173] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquiring lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.256556] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.256778] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquiring lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.256967] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.260158] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.260158] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.260158] env[63293]: value = "task-1327572" [ 822.260158] env[63293]: _type = "Task" [ 822.260158] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.260557] env[63293]: INFO nova.compute.manager [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Terminating instance [ 822.273075] env[63293]: DEBUG nova.compute.manager [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 822.273336] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 822.274413] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715d85d2-f1fa-4d9d-8df8-48bdc859fea1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.283928] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327572, 'name': CreateVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.286545] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 822.286836] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd123bc3-6f76-4a8b-a731-58b72108ffc3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.295109] env[63293]: DEBUG oslo_vmware.api [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for the task: (returnval){ [ 822.295109] env[63293]: value = "task-1327573" [ 822.295109] env[63293]: _type = "Task" [ 822.295109] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.304548] env[63293]: DEBUG oslo_vmware.api [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327573, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.582236] env[63293]: DEBUG nova.compute.utils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.589083] env[63293]: DEBUG nova.compute.manager [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.589083] env[63293]: DEBUG nova.network.neutron [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 822.667661] env[63293]: DEBUG nova.policy [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ecbac9f9fca4b42a14c3ca0d26e5d34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'affba61d6a2846b38666544bc2c25db5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.673685] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327566, 'name': ReconfigVM_Task, 'duration_secs': 1.098401} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.673685] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 33d71260-26f1-482c-b93b-5f8e27c545f3/33d71260-26f1-482c-b93b-5f8e27c545f3.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.673685] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ab40403-b021-4d4b-abf2-924cffa97c64 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.687178] env[63293]: DEBUG oslo_vmware.api [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Task: {'id': task-1327568, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148801} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.688501] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 822.688698] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 822.688886] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 822.689102] env[63293]: INFO nova.compute.manager [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Took 1.66 seconds to destroy the instance on the hypervisor. [ 822.689342] env[63293]: DEBUG oslo.service.loopingcall [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.689600] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 822.689600] env[63293]: value = "task-1327574" [ 822.689600] env[63293]: _type = "Task" [ 822.689600] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.689787] env[63293]: DEBUG nova.compute.manager [-] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.689896] env[63293]: DEBUG nova.network.neutron [-] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 822.703891] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327574, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.747181] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327570, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.748748] env[63293]: INFO nova.compute.manager [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Took 44.48 seconds to build instance. [ 822.774768] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327572, 'name': CreateVM_Task, 'duration_secs': 0.359329} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.774965] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 822.775711] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283694', 'volume_id': '1f5a039e-f604-47d3-a580-f1754a5b697d', 'name': 'volume-1f5a039e-f604-47d3-a580-f1754a5b697d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8637881-9fb6-4c5d-848d-7b2d38f8e970', 'attached_at': '', 'detached_at': '', 'volume_id': '1f5a039e-f604-47d3-a580-f1754a5b697d', 'serial': '1f5a039e-f604-47d3-a580-f1754a5b697d'}, 'attachment_id': 'bbfde594-d781-4afc-8141-496cc4f49ab2', 'device_type': None, 'delete_on_termination': True, 'mount_device': '/dev/sda', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=63293) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 822.775881] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Root volume attach. Driver type: vmdk {{(pid=63293) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 822.777687] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88710ae-fd8a-4061-a4ff-e53bed0d6979 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.792541] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c4065f-7349-4b65-b0d9-b5fbd7600739 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.809782] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448d675b-20ab-42d6-aa62-34577ea621e5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.812749] env[63293]: DEBUG oslo_vmware.api [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327573, 'name': PowerOffVM_Task, 'duration_secs': 0.200967} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.813103] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 822.813215] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 822.813755] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9bbd021-a93d-4539-88bb-6add683a4dfa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.819766] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-299a19bb-208d-4e3b-a1c0-7baaf27ca15c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.826578] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for the task: (returnval){ [ 822.826578] env[63293]: value = "task-1327576" [ 822.826578] env[63293]: _type = "Task" [ 822.826578] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.838293] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327576, 'name': RelocateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.881560] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 822.881560] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 822.881560] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Deleting the datastore file [datastore1] c8009fc0-f7dc-4568-9909-7fa9f7f6a56c {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 822.882091] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-24349ad3-8d99-439f-89e4-d3ee6c0ccc8e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.888559] env[63293]: DEBUG oslo_vmware.api [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for the task: (returnval){ [ 822.888559] env[63293]: value = "task-1327577" [ 822.888559] env[63293]: _type = "Task" [ 822.888559] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.895106] env[63293]: DEBUG nova.compute.manager [req-4386046d-0457-4130-b36a-7fad05a75c43 req-650580ff-46ce-43e5-b6c5-a522a19a02c8 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Received event network-vif-plugged-26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.895106] env[63293]: DEBUG oslo_concurrency.lockutils [req-4386046d-0457-4130-b36a-7fad05a75c43 req-650580ff-46ce-43e5-b6c5-a522a19a02c8 service nova] Acquiring lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.895106] env[63293]: DEBUG oslo_concurrency.lockutils [req-4386046d-0457-4130-b36a-7fad05a75c43 req-650580ff-46ce-43e5-b6c5-a522a19a02c8 service nova] Lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.895106] env[63293]: DEBUG oslo_concurrency.lockutils [req-4386046d-0457-4130-b36a-7fad05a75c43 req-650580ff-46ce-43e5-b6c5-a522a19a02c8 service nova] Lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.895319] env[63293]: DEBUG nova.compute.manager [req-4386046d-0457-4130-b36a-7fad05a75c43 req-650580ff-46ce-43e5-b6c5-a522a19a02c8 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] No waiting events found dispatching network-vif-plugged-26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 822.895356] env[63293]: WARNING nova.compute.manager [req-4386046d-0457-4130-b36a-7fad05a75c43 req-650580ff-46ce-43e5-b6c5-a522a19a02c8 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Received unexpected event network-vif-plugged-26b86f34-f9b6-4b0b-abdd-d8acd51085fd for instance with vm_state building and task_state spawning. [ 822.903725] env[63293]: DEBUG oslo_vmware.api [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327577, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.972574] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf298ca5-811a-456f-83fe-e9810edac51c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.981469] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0da0e61-8738-49d3-803f-3c5b68aed302 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.013606] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f575e62-1d79-4bbe-9c9f-bbf9e5735e14 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.021633] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04af9b55-6bd2-469d-9499-a22fbd1db5c9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.039208] env[63293]: DEBUG nova.compute.provider_tree [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.087688] env[63293]: DEBUG nova.compute.manager [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 823.203566] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327574, 'name': Rename_Task, 'duration_secs': 0.135851} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.203853] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.204131] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1216ebcb-85fb-4d9e-9ee6-583a9596d15c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.214090] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 823.214090] env[63293]: value = "task-1327578" [ 823.214090] env[63293]: _type = "Task" [ 823.214090] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.226497] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327578, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.236482] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327570, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.251383] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd217318-df58-4fac-810e-b85b645dd376 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.901s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.341973] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327576, 'name': RelocateVM_Task} progress is 42%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.401752] env[63293]: DEBUG nova.network.neutron [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Successfully created port: 0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.412956] env[63293]: DEBUG oslo_vmware.api [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Task: {'id': task-1327577, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138753} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.413344] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.413522] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 823.413731] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.415696] env[63293]: INFO nova.compute.manager [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Took 1.14 seconds to destroy the instance on the hypervisor. [ 823.416011] env[63293]: DEBUG oslo.service.loopingcall [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.416234] env[63293]: DEBUG nova.compute.manager [-] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 823.416327] env[63293]: DEBUG nova.network.neutron [-] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 823.509190] env[63293]: DEBUG nova.network.neutron [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Successfully updated port: 26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 823.545152] env[63293]: DEBUG nova.scheduler.client.report [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.559576] env[63293]: DEBUG nova.network.neutron [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Updated VIF entry in instance network info cache for port aa552093-dd1a-4dd1-aadf-9c92b766b4ea. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 823.559968] env[63293]: DEBUG nova.network.neutron [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Updating instance_info_cache with network_info: [{"id": "aa552093-dd1a-4dd1-aadf-9c92b766b4ea", "address": "fa:16:3e:13:f8:5d", "network": {"id": "0ec4e867-7095-4c29-9202-669ad2701e74", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-378057015-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fd63e6892c44f159195897ad5eacf13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa552093-dd", "ovs_interfaceid": "aa552093-dd1a-4dd1-aadf-9c92b766b4ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.727035] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327578, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.737795] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327570, 'name': CloneVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.754929] env[63293]: DEBUG nova.compute.manager [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 823.829266] env[63293]: DEBUG nova.network.neutron [-] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.846343] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327576, 'name': RelocateVM_Task} progress is 54%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.013466] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquiring lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.013510] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquired lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.013637] env[63293]: DEBUG nova.network.neutron [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.057022] env[63293]: DEBUG oslo_concurrency.lockutils [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.977s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.062393] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.126s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.067025] env[63293]: INFO nova.compute.claims [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.069576] env[63293]: DEBUG oslo_concurrency.lockutils [req-91393e64-fc0b-48fe-bc4b-05dcc8ca5977 req-b9c69170-2d49-4d51-9f10-c0259fedb5d7 service nova] Releasing lock "refresh_cache-b8637881-9fb6-4c5d-848d-7b2d38f8e970" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.100188] env[63293]: DEBUG nova.compute.manager [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 824.104759] env[63293]: INFO nova.scheduler.client.report [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Deleted allocations for instance b48324a4-902d-4642-8827-daee9683b3a0 [ 824.132579] env[63293]: DEBUG nova.virt.hardware [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.132729] env[63293]: DEBUG nova.virt.hardware [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.132905] env[63293]: DEBUG nova.virt.hardware [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.133078] env[63293]: DEBUG nova.virt.hardware [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.133235] env[63293]: DEBUG nova.virt.hardware [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.134518] env[63293]: DEBUG nova.virt.hardware [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.134518] env[63293]: DEBUG nova.virt.hardware [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.134518] env[63293]: DEBUG nova.virt.hardware [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.134518] env[63293]: DEBUG nova.virt.hardware [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.134518] env[63293]: DEBUG nova.virt.hardware [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.135413] env[63293]: DEBUG nova.virt.hardware [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.135473] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45240bd8-3070-4f26-872d-c8acb547c91f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.146836] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4faf9f4c-0c5a-42fe-ade1-67256e9b5638 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.244507] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327578, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.257297] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327570, 'name': CloneVM_Task, 'duration_secs': 1.575676} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.258621] env[63293]: INFO nova.virt.vmwareapi.vmops [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Created linked-clone VM from snapshot [ 824.259064] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609cc462-5ccc-4b79-84ca-86cdf9933234 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.279817] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Uploading image 54404016-2794-4f2d-a4d4-8ec6658c7f45 {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 824.287942] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.313932] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 824.313932] env[63293]: value = "vm-283725" [ 824.313932] env[63293]: _type = "VirtualMachine" [ 824.313932] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 824.315506] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c4c6a69e-7664-4003-8be0-b310928b9d77 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.327981] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lease: (returnval){ [ 824.327981] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527e89ca-fba3-5107-dc39-996ba04afed5" [ 824.327981] env[63293]: _type = "HttpNfcLease" [ 824.327981] env[63293]: } obtained for exporting VM: (result){ [ 824.327981] env[63293]: value = "vm-283725" [ 824.327981] env[63293]: _type = "VirtualMachine" [ 824.327981] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 824.328895] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the lease: (returnval){ [ 824.328895] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527e89ca-fba3-5107-dc39-996ba04afed5" [ 824.328895] env[63293]: _type = "HttpNfcLease" [ 824.328895] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 824.337716] env[63293]: INFO nova.compute.manager [-] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Took 1.65 seconds to deallocate network for instance. [ 824.347189] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327576, 'name': RelocateVM_Task} progress is 67%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.347442] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 824.347442] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527e89ca-fba3-5107-dc39-996ba04afed5" [ 824.347442] env[63293]: _type = "HttpNfcLease" [ 824.347442] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 824.347681] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 824.347681] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527e89ca-fba3-5107-dc39-996ba04afed5" [ 824.347681] env[63293]: _type = "HttpNfcLease" [ 824.347681] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 824.348454] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf03476b-db60-46fd-aa6b-06c36a255b03 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.361417] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521315ef-5328-4c62-2280-221bd775fa1e/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 824.361417] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521315ef-5328-4c62-2280-221bd775fa1e/disk-0.vmdk for reading. {{(pid=63293) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 824.481630] env[63293]: DEBUG nova.compute.manager [req-9b4a2761-e6c4-41a2-8c9a-c38e1cfdcaa1 req-709f6cab-e49a-4600-8522-8550f1196ee8 service nova] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Received event network-vif-deleted-70383439-e38d-4f62-8f53-350c7b430342 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.481630] env[63293]: DEBUG nova.compute.manager [req-9b4a2761-e6c4-41a2-8c9a-c38e1cfdcaa1 req-709f6cab-e49a-4600-8522-8550f1196ee8 service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Received event network-vif-deleted-e35e7768-ec67-4322-a4c5-be6a5835054e {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.481733] env[63293]: INFO nova.compute.manager [req-9b4a2761-e6c4-41a2-8c9a-c38e1cfdcaa1 req-709f6cab-e49a-4600-8522-8550f1196ee8 service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Neutron deleted interface e35e7768-ec67-4322-a4c5-be6a5835054e; detaching it from the instance and deleting it from the info cache [ 824.481889] env[63293]: DEBUG nova.network.neutron [req-9b4a2761-e6c4-41a2-8c9a-c38e1cfdcaa1 req-709f6cab-e49a-4600-8522-8550f1196ee8 service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.493901] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ce6518eb-003c-4c92-9af7-8233d7bbd75f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.564225] env[63293]: DEBUG nova.network.neutron [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.618654] env[63293]: DEBUG oslo_concurrency.lockutils [None req-297b4309-993f-48aa-89fe-b0590cef8923 tempest-VolumesAdminNegativeTest-932678640 tempest-VolumesAdminNegativeTest-932678640-project-member] Lock "b48324a4-902d-4642-8827-daee9683b3a0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.297s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.705970] env[63293]: DEBUG nova.network.neutron [-] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.734093] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327578, 'name': PowerOnVM_Task} progress is 76%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.768035] env[63293]: DEBUG nova.network.neutron [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updating instance_info_cache with network_info: [{"id": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "address": "fa:16:3e:fc:c1:7d", "network": {"id": "bbf31454-8e46-45d9-b10a-fccda17654c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1875587777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "6fa6cd8dea16412db105faf2ff93a6ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26b86f34-f9", "ovs_interfaceid": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.844321] env[63293]: DEBUG oslo_concurrency.lockutils [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.851767] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327576, 'name': RelocateVM_Task} progress is 82%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.924139] env[63293]: DEBUG nova.compute.manager [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Received event network-changed-26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.924597] env[63293]: DEBUG nova.compute.manager [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Refreshing instance network info cache due to event network-changed-26b86f34-f9b6-4b0b-abdd-d8acd51085fd. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 824.924830] env[63293]: DEBUG oslo_concurrency.lockutils [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] Acquiring lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.987809] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff4fb2b4-fcf5-4127-b7b3-1c52c03fb355 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.002881] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58670c7-425e-4973-88ce-8f268abeec07 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.034455] env[63293]: DEBUG nova.compute.manager [req-9b4a2761-e6c4-41a2-8c9a-c38e1cfdcaa1 req-709f6cab-e49a-4600-8522-8550f1196ee8 service nova] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Detach interface failed, port_id=e35e7768-ec67-4322-a4c5-be6a5835054e, reason: Instance c8009fc0-f7dc-4568-9909-7fa9f7f6a56c could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 825.082154] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquiring lock "828a1165-3829-431c-9edb-3a5d3a8054ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.082589] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "828a1165-3829-431c-9edb-3a5d3a8054ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.214295] env[63293]: INFO nova.compute.manager [-] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Took 1.80 seconds to deallocate network for instance. [ 825.238367] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327578, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.270749] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Releasing lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.271136] env[63293]: DEBUG nova.compute.manager [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Instance network_info: |[{"id": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "address": "fa:16:3e:fc:c1:7d", "network": {"id": "bbf31454-8e46-45d9-b10a-fccda17654c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1875587777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "6fa6cd8dea16412db105faf2ff93a6ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26b86f34-f9", "ovs_interfaceid": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 825.271740] env[63293]: DEBUG oslo_concurrency.lockutils [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] Acquired lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.271974] env[63293]: DEBUG nova.network.neutron [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Refreshing network info cache for port 26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.273604] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:c1:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271a82f1-1d09-4ad3-9c15-07269bad114c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '26b86f34-f9b6-4b0b-abdd-d8acd51085fd', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.283717] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Creating folder: Project (6fa6cd8dea16412db105faf2ff93a6ac). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.288296] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e05ceabf-16bf-47d9-8381-2e451de0494e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.301031] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Created folder: Project (6fa6cd8dea16412db105faf2ff93a6ac) in parent group-v283678. [ 825.303786] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Creating folder: Instances. Parent ref: group-v283727. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.303786] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f83dd83-575e-440c-81cf-d0d63ab524e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.314513] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Created folder: Instances in parent group-v283727. [ 825.314513] env[63293]: DEBUG oslo.service.loopingcall [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.314876] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.315192] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c569e99-55a1-4386-ac33-a0a3df7147d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.342312] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.342312] env[63293]: value = "task-1327582" [ 825.342312] env[63293]: _type = "Task" [ 825.342312] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.346783] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327576, 'name': RelocateVM_Task} progress is 97%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.358277] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327582, 'name': CreateVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.493652] env[63293]: DEBUG nova.compute.manager [req-a1c8142d-f533-4246-a9f3-7b2b17e8bf70 req-312543d0-67cd-4ccf-b12b-6f29992190b3 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Received event network-vif-plugged-0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.493974] env[63293]: DEBUG oslo_concurrency.lockutils [req-a1c8142d-f533-4246-a9f3-7b2b17e8bf70 req-312543d0-67cd-4ccf-b12b-6f29992190b3 service nova] Acquiring lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.494395] env[63293]: DEBUG oslo_concurrency.lockutils [req-a1c8142d-f533-4246-a9f3-7b2b17e8bf70 req-312543d0-67cd-4ccf-b12b-6f29992190b3 service nova] Lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.494570] env[63293]: DEBUG oslo_concurrency.lockutils [req-a1c8142d-f533-4246-a9f3-7b2b17e8bf70 req-312543d0-67cd-4ccf-b12b-6f29992190b3 service nova] Lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.494801] env[63293]: DEBUG nova.compute.manager [req-a1c8142d-f533-4246-a9f3-7b2b17e8bf70 req-312543d0-67cd-4ccf-b12b-6f29992190b3 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] No waiting events found dispatching network-vif-plugged-0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.495072] env[63293]: WARNING nova.compute.manager [req-a1c8142d-f533-4246-a9f3-7b2b17e8bf70 req-312543d0-67cd-4ccf-b12b-6f29992190b3 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Received unexpected event network-vif-plugged-0576b708-1b9b-4159-8079-9b97d8b4e70d for instance with vm_state building and task_state spawning. [ 825.501132] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f5ce21-204d-41d3-8f25-d313eef90ce6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.514307] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c48b075-497f-4e75-ad70-4b783b63545a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.564806] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263416df-80c6-4142-b1d8-e708b3e8577b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.573977] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581bce8b-cb06-4a05-bfd6-e399953cbc86 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.591579] env[63293]: DEBUG nova.compute.provider_tree [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.732548] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.733170] env[63293]: DEBUG oslo_vmware.api [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327578, 'name': PowerOnVM_Task, 'duration_secs': 2.29339} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.735053] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.736600] env[63293]: INFO nova.compute.manager [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Took 10.55 seconds to spawn the instance on the hypervisor. [ 825.736600] env[63293]: DEBUG nova.compute.manager [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.737238] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29386dae-b157-4630-8caf-b2bddd0c3350 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.845923] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327576, 'name': RelocateVM_Task} progress is 97%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.853865] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327582, 'name': CreateVM_Task, 'duration_secs': 0.399423} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.854037] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.854713] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.854874] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.855214] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.855463] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25bc3060-79ac-409f-ba3c-569a138930cb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.860074] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 825.860074] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523fbfa8-73d6-831b-a558-f97438b9741d" [ 825.860074] env[63293]: _type = "Task" [ 825.860074] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.869323] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523fbfa8-73d6-831b-a558-f97438b9741d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.038355] env[63293]: DEBUG nova.network.neutron [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updated VIF entry in instance network info cache for port 26b86f34-f9b6-4b0b-abdd-d8acd51085fd. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 826.038731] env[63293]: DEBUG nova.network.neutron [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updating instance_info_cache with network_info: [{"id": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "address": "fa:16:3e:fc:c1:7d", "network": {"id": "bbf31454-8e46-45d9-b10a-fccda17654c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1875587777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "6fa6cd8dea16412db105faf2ff93a6ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26b86f34-f9", "ovs_interfaceid": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.049884] env[63293]: DEBUG nova.network.neutron [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Successfully updated port: 0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 826.094713] env[63293]: DEBUG nova.scheduler.client.report [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.264888] env[63293]: INFO nova.compute.manager [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Took 46.03 seconds to build instance. [ 826.347465] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327576, 'name': RelocateVM_Task} progress is 97%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.370995] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523fbfa8-73d6-831b-a558-f97438b9741d, 'name': SearchDatastore_Task, 'duration_secs': 0.009658} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.371369] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.371626] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.371872] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.372036] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.372226] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.372504] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-25b9b28f-aa05-4135-8b23-8c4142c558a2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.381529] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.381725] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 826.382539] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80bd4df1-e5bf-421e-8962-6db865ac5b6b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.388073] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 826.388073] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d6a59d-6bda-8653-1ed1-a588e9cc0b42" [ 826.388073] env[63293]: _type = "Task" [ 826.388073] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.398734] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d6a59d-6bda-8653-1ed1-a588e9cc0b42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.509145] env[63293]: DEBUG nova.compute.manager [req-ed683ad5-f966-4a4a-9d48-47d797c55c2c req-3e06c381-c09a-47d8-a024-062e4b2ead0c service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Received event network-changed-0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 826.509404] env[63293]: DEBUG nova.compute.manager [req-ed683ad5-f966-4a4a-9d48-47d797c55c2c req-3e06c381-c09a-47d8-a024-062e4b2ead0c service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Refreshing instance network info cache due to event network-changed-0576b708-1b9b-4159-8079-9b97d8b4e70d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 826.509684] env[63293]: DEBUG oslo_concurrency.lockutils [req-ed683ad5-f966-4a4a-9d48-47d797c55c2c req-3e06c381-c09a-47d8-a024-062e4b2ead0c service nova] Acquiring lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.509865] env[63293]: DEBUG oslo_concurrency.lockutils [req-ed683ad5-f966-4a4a-9d48-47d797c55c2c req-3e06c381-c09a-47d8-a024-062e4b2ead0c service nova] Acquired lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.510111] env[63293]: DEBUG nova.network.neutron [req-ed683ad5-f966-4a4a-9d48-47d797c55c2c req-3e06c381-c09a-47d8-a024-062e4b2ead0c service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Refreshing network info cache for port 0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 826.541895] env[63293]: DEBUG oslo_concurrency.lockutils [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] Releasing lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.543135] env[63293]: DEBUG nova.compute.manager [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received event network-changed-a34b2cc4-afe1-463c-bb22-470ea54d8789 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 826.543135] env[63293]: DEBUG nova.compute.manager [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Refreshing instance network info cache due to event network-changed-a34b2cc4-afe1-463c-bb22-470ea54d8789. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 826.543548] env[63293]: DEBUG oslo_concurrency.lockutils [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] Acquiring lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.543831] env[63293]: DEBUG oslo_concurrency.lockutils [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] Acquired lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.544136] env[63293]: DEBUG nova.network.neutron [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Refreshing network info cache for port a34b2cc4-afe1-463c-bb22-470ea54d8789 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 826.552674] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.599693] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.601129] env[63293]: DEBUG nova.compute.manager [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.604392] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.676s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.606277] env[63293]: INFO nova.compute.claims [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.767603] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e55fd9d7-134f-4987-9891-b589811cccec tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "33d71260-26f1-482c-b93b-5f8e27c545f3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.053s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.852429] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327576, 'name': RelocateVM_Task, 'duration_secs': 4.020316} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.853201] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Volume attach. Driver type: vmdk {{(pid=63293) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 826.853506] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283694', 'volume_id': '1f5a039e-f604-47d3-a580-f1754a5b697d', 'name': 'volume-1f5a039e-f604-47d3-a580-f1754a5b697d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8637881-9fb6-4c5d-848d-7b2d38f8e970', 'attached_at': '', 'detached_at': '', 'volume_id': '1f5a039e-f604-47d3-a580-f1754a5b697d', 'serial': '1f5a039e-f604-47d3-a580-f1754a5b697d'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 826.854327] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ceabfe8-d39a-4306-a662-51021860e9e8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.878104] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91cc838-1c34-4dd5-98ee-1f41625e1cef {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.903659] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] volume-1f5a039e-f604-47d3-a580-f1754a5b697d/volume-1f5a039e-f604-47d3-a580-f1754a5b697d.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.909105] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6869774d-79f1-41c6-82f1-0dfb910187bb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.932723] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d6a59d-6bda-8653-1ed1-a588e9cc0b42, 'name': SearchDatastore_Task, 'duration_secs': 0.010224} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.934794] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for the task: (returnval){ [ 826.934794] env[63293]: value = "task-1327583" [ 826.934794] env[63293]: _type = "Task" [ 826.934794] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.935033] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f727ba59-9c7c-4738-a01b-8b5b4e541bfb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.943017] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 826.943017] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527cc63e-b23c-7fd7-5daf-d3cc96b0b245" [ 826.943017] env[63293]: _type = "Task" [ 826.943017] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.946743] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327583, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.954336] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527cc63e-b23c-7fd7-5daf-d3cc96b0b245, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.050910] env[63293]: DEBUG nova.network.neutron [req-ed683ad5-f966-4a4a-9d48-47d797c55c2c req-3e06c381-c09a-47d8-a024-062e4b2ead0c service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.115830] env[63293]: DEBUG nova.compute.utils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.119702] env[63293]: DEBUG nova.compute.manager [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.119760] env[63293]: DEBUG nova.network.neutron [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 827.197807] env[63293]: DEBUG nova.policy [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '21199702ac8a410191472977f096be81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76803a169e7d424ab6aefe0d772b7827', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 827.260887] env[63293]: DEBUG nova.network.neutron [req-ed683ad5-f966-4a4a-9d48-47d797c55c2c req-3e06c381-c09a-47d8-a024-062e4b2ead0c service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.271276] env[63293]: DEBUG nova.compute.manager [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 827.452582] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327583, 'name': ReconfigVM_Task, 'duration_secs': 0.282037} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.460515] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Reconfigured VM instance instance-0000003d to attach disk [datastore2] volume-1f5a039e-f604-47d3-a580-f1754a5b697d/volume-1f5a039e-f604-47d3-a580-f1754a5b697d.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.463923] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74fd0994-bf6b-419d-aded-2ce613cc425d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.482288] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527cc63e-b23c-7fd7-5daf-d3cc96b0b245, 'name': SearchDatastore_Task, 'duration_secs': 0.012757} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.485037] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.485037] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 48603812-c3cc-4dae-b8ba-99b9ac5f7969/48603812-c3cc-4dae-b8ba-99b9ac5f7969.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.485037] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for the task: (returnval){ [ 827.485037] env[63293]: value = "task-1327584" [ 827.485037] env[63293]: _type = "Task" [ 827.485037] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.485037] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f10070a8-216a-4dac-83bb-9e8b9500c8e8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.495909] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327584, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.498341] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 827.498341] env[63293]: value = "task-1327585" [ 827.498341] env[63293]: _type = "Task" [ 827.498341] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.511533] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327585, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.588300] env[63293]: DEBUG nova.network.neutron [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Successfully created port: 52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.603356] env[63293]: DEBUG nova.network.neutron [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updated VIF entry in instance network info cache for port a34b2cc4-afe1-463c-bb22-470ea54d8789. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 827.604556] env[63293]: DEBUG nova.network.neutron [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [{"id": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "address": "fa:16:3e:2b:1c:64", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b2cc4-af", "ovs_interfaceid": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.624285] env[63293]: DEBUG nova.compute.manager [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.763963] env[63293]: DEBUG oslo_concurrency.lockutils [req-ed683ad5-f966-4a4a-9d48-47d797c55c2c req-3e06c381-c09a-47d8-a024-062e4b2ead0c service nova] Releasing lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.764552] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.764886] env[63293]: DEBUG nova.network.neutron [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.795445] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.002635] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327584, 'name': ReconfigVM_Task, 'duration_secs': 0.177747} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.010203] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283694', 'volume_id': '1f5a039e-f604-47d3-a580-f1754a5b697d', 'name': 'volume-1f5a039e-f604-47d3-a580-f1754a5b697d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8637881-9fb6-4c5d-848d-7b2d38f8e970', 'attached_at': '', 'detached_at': '', 'volume_id': '1f5a039e-f604-47d3-a580-f1754a5b697d', 'serial': '1f5a039e-f604-47d3-a580-f1754a5b697d'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 828.011448] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85e66a6d-c9a8-4ded-be35-56bcb608b38d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.021317] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327585, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.023864] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for the task: (returnval){ [ 828.023864] env[63293]: value = "task-1327586" [ 828.023864] env[63293]: _type = "Task" [ 828.023864] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.038118] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327586, 'name': Rename_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.039438] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23376205-7929-4542-a851-64964c7b0474 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.048410] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fd5952-0d08-4521-9fee-b942ef1cc5e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.084452] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee58982-6b59-40b4-b1a2-9d4d5b4cf1be {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.094436] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ed9ea9-8560-4357-abd8-810eb36ea625 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.112911] env[63293]: DEBUG oslo_concurrency.lockutils [req-a310b580-5e56-4bbc-85d4-c60a07a228e1 req-cd2aeeb4-ff52-4679-b475-968656e01001 service nova] Releasing lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.113702] env[63293]: DEBUG nova.compute.provider_tree [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.339595] env[63293]: DEBUG nova.network.neutron [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.417803] env[63293]: INFO nova.compute.manager [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Rebuilding instance [ 828.468219] env[63293]: DEBUG nova.compute.manager [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.469120] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c3482b-8d8f-4f82-98d3-471ca52697a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.512103] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327585, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.769067} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.512396] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 48603812-c3cc-4dae-b8ba-99b9ac5f7969/48603812-c3cc-4dae-b8ba-99b9ac5f7969.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.512616] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.512870] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc1112be-fe9c-4ae1-9a6d-1dba519a1e10 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.519489] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 828.519489] env[63293]: value = "task-1327587" [ 828.519489] env[63293]: _type = "Task" [ 828.519489] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.528095] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.537159] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327586, 'name': Rename_Task, 'duration_secs': 0.372648} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.537358] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 828.537605] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce1da2ba-4a43-4e3b-8cdf-d4cfc20f397d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.543345] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for the task: (returnval){ [ 828.543345] env[63293]: value = "task-1327588" [ 828.543345] env[63293]: _type = "Task" [ 828.543345] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.551010] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327588, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.617662] env[63293]: DEBUG nova.scheduler.client.report [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.640966] env[63293]: DEBUG nova.compute.manager [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.651420] env[63293]: DEBUG nova.network.neutron [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updating instance_info_cache with network_info: [{"id": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "address": "fa:16:3e:c0:e8:15", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0576b708-1b", "ovs_interfaceid": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.672222] env[63293]: DEBUG nova.virt.hardware [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.672517] env[63293]: DEBUG nova.virt.hardware [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.672690] env[63293]: DEBUG nova.virt.hardware [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.672908] env[63293]: DEBUG nova.virt.hardware [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.673098] env[63293]: DEBUG nova.virt.hardware [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.673282] env[63293]: DEBUG nova.virt.hardware [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.673519] env[63293]: DEBUG nova.virt.hardware [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.673709] env[63293]: DEBUG nova.virt.hardware [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.673889] env[63293]: DEBUG nova.virt.hardware [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.674071] env[63293]: DEBUG nova.virt.hardware [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.674254] env[63293]: DEBUG nova.virt.hardware [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.675852] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb378ea3-b888-4f5d-87e2-4f007ff33fc5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.685030] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec02736-2cbb-4311-8846-e592c918a7d6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.980755] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 828.981572] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7622ca6a-f7e2-4f33-835e-5411925a723a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.991017] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 828.991017] env[63293]: value = "task-1327589" [ 828.991017] env[63293]: _type = "Task" [ 828.991017] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.000219] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327589, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.030190] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.138887} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.030521] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.031353] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95cd9772-65b3-4be0-9c8e-70f3c36dbd34 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.057493] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 48603812-c3cc-4dae-b8ba-99b9ac5f7969/48603812-c3cc-4dae-b8ba-99b9ac5f7969.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.061305] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-513fc2a9-34c6-4ae5-9504-be2d9fbd3de9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.085361] env[63293]: DEBUG oslo_vmware.api [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327588, 'name': PowerOnVM_Task, 'duration_secs': 0.502007} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.086780] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 829.087015] env[63293]: INFO nova.compute.manager [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Took 7.73 seconds to spawn the instance on the hypervisor. [ 829.087214] env[63293]: DEBUG nova.compute.manager [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.087574] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 829.087574] env[63293]: value = "task-1327590" [ 829.087574] env[63293]: _type = "Task" [ 829.087574] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.088345] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6d71af-4747-4a20-a799-4d9d69d6d4c5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.109388] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.124537] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.125162] env[63293]: DEBUG nova.compute.manager [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.129777] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.072s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.136025] env[63293]: INFO nova.compute.claims [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.154233] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.154233] env[63293]: DEBUG nova.compute.manager [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Instance network_info: |[{"id": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "address": "fa:16:3e:c0:e8:15", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0576b708-1b", "ovs_interfaceid": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 829.156676] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:e8:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24144f5a-050a-4f1e-8d8c-774dc16dc791', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0576b708-1b9b-4159-8079-9b97d8b4e70d', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.163849] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating folder: Project (affba61d6a2846b38666544bc2c25db5). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.165687] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a3434776-5df9-44c5-bc18-e5d68cf2aaba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.174975] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Created folder: Project (affba61d6a2846b38666544bc2c25db5) in parent group-v283678. [ 829.175117] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating folder: Instances. Parent ref: group-v283730. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.175420] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce03ac64-dbb3-41e3-9e49-1bcfe809e2da {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.185530] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Created folder: Instances in parent group-v283730. [ 829.185817] env[63293]: DEBUG oslo.service.loopingcall [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.186065] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 829.186319] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c60f3921-3ff1-41a9-aa76-b6b5a10ef1cd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.207271] env[63293]: DEBUG nova.compute.manager [req-236979d1-727e-416e-b043-060b5c365060 req-8f159ebd-a755-4323-91c1-c1b0459e0ca2 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Received event network-vif-plugged-52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.207684] env[63293]: DEBUG oslo_concurrency.lockutils [req-236979d1-727e-416e-b043-060b5c365060 req-8f159ebd-a755-4323-91c1-c1b0459e0ca2 service nova] Acquiring lock "831d7d9b-364f-4c29-bda5-e6c38291a973-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.207991] env[63293]: DEBUG oslo_concurrency.lockutils [req-236979d1-727e-416e-b043-060b5c365060 req-8f159ebd-a755-4323-91c1-c1b0459e0ca2 service nova] Lock "831d7d9b-364f-4c29-bda5-e6c38291a973-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.208144] env[63293]: DEBUG oslo_concurrency.lockutils [req-236979d1-727e-416e-b043-060b5c365060 req-8f159ebd-a755-4323-91c1-c1b0459e0ca2 service nova] Lock "831d7d9b-364f-4c29-bda5-e6c38291a973-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.208300] env[63293]: DEBUG nova.compute.manager [req-236979d1-727e-416e-b043-060b5c365060 req-8f159ebd-a755-4323-91c1-c1b0459e0ca2 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] No waiting events found dispatching network-vif-plugged-52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 829.208530] env[63293]: WARNING nova.compute.manager [req-236979d1-727e-416e-b043-060b5c365060 req-8f159ebd-a755-4323-91c1-c1b0459e0ca2 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Received unexpected event network-vif-plugged-52a5b959-25ef-48ed-ad04-82f85a131929 for instance with vm_state building and task_state spawning. [ 829.214677] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.214677] env[63293]: value = "task-1327593" [ 829.214677] env[63293]: _type = "Task" [ 829.214677] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.225317] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327593, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.250389] env[63293]: DEBUG nova.network.neutron [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Successfully updated port: 52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.504186] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327589, 'name': PowerOffVM_Task, 'duration_secs': 0.218705} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.504572] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 829.505028] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.506152] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f065cec3-1332-4ce2-a2cb-f2a022683048 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.517613] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 829.517923] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80569ac2-e00d-471f-8218-f3493341a86d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.599702] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 829.600261] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 829.600648] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleting the datastore file [datastore1] 33d71260-26f1-482c-b93b-5f8e27c545f3 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.609829] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c547bc3b-68b9-400e-abef-442ee0a3111f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.622759] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.624272] env[63293]: INFO nova.compute.manager [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Took 44.65 seconds to build instance. [ 829.625665] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 829.625665] env[63293]: value = "task-1327595" [ 829.625665] env[63293]: _type = "Task" [ 829.625665] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.637895] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327595, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.638455] env[63293]: DEBUG nova.compute.utils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.644240] env[63293]: DEBUG nova.compute.manager [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.644240] env[63293]: DEBUG nova.network.neutron [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 829.725800] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327593, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.754431] env[63293]: DEBUG nova.policy [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a3094206a5324d3ea72b18ac38fd3afa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b461ff81505c49dbbd9eb64935492b43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.754664] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquiring lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.757043] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquired lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.757043] env[63293]: DEBUG nova.network.neutron [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 830.110867] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327590, 'name': ReconfigVM_Task, 'duration_secs': 0.615173} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.111219] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 48603812-c3cc-4dae-b8ba-99b9ac5f7969/48603812-c3cc-4dae-b8ba-99b9ac5f7969.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.111879] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6fed43f5-3240-4d9f-a356-85bb3d6c9de8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.119159] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 830.119159] env[63293]: value = "task-1327596" [ 830.119159] env[63293]: _type = "Task" [ 830.119159] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.128494] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9d1b935c-7e6a-43c1-9fca-dfa4491a1654 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.632s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.128686] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327596, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.143297] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327595, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.237815} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.147146] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.147395] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 830.149407] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 830.151243] env[63293]: DEBUG nova.compute.manager [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.205633] env[63293]: DEBUG nova.network.neutron [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Successfully created port: 1582d90f-11d5-4ec4-8173-e1fcb11d3bc8 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.227676] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327593, 'name': CreateVM_Task, 'duration_secs': 0.701425} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.231360] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 830.232652] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.232790] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.233173] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.233442] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4dc6b727-2265-4a5a-b51e-1111b0be3417 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.245838] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 830.245838] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529fb8c3-912c-b6e6-9a2f-b8cdd5da87b3" [ 830.245838] env[63293]: _type = "Task" [ 830.245838] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.256541] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529fb8c3-912c-b6e6-9a2f-b8cdd5da87b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.317792] env[63293]: DEBUG nova.network.neutron [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.500547] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e47c3fb-96c4-445f-89ad-39c61e720e7a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.509800] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69dba2b1-ae78-4511-9020-6b023671d392 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.542987] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157102fd-2864-46cc-ab55-7aaa18f0afe7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.551177] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ead601-e372-4b4c-a8e7-403d159e8355 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.566157] env[63293]: DEBUG nova.compute.provider_tree [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.643486] env[63293]: DEBUG nova.compute.manager [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 830.646235] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327596, 'name': Rename_Task, 'duration_secs': 0.177954} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.646235] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 830.646235] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-940f3584-54e0-4b66-a5f9-dabec2689380 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.668353] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 830.668353] env[63293]: value = "task-1327597" [ 830.668353] env[63293]: _type = "Task" [ 830.668353] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.681076] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327597, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.684668] env[63293]: DEBUG nova.network.neutron [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updating instance_info_cache with network_info: [{"id": "52a5b959-25ef-48ed-ad04-82f85a131929", "address": "fa:16:3e:56:4e:4a", "network": {"id": "b6024f78-d9d9-4fd8-a71e-9f996ce4a30f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-959509061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76803a169e7d424ab6aefe0d772b7827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52a5b959-25", "ovs_interfaceid": "52a5b959-25ef-48ed-ad04-82f85a131929", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.756159] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529fb8c3-912c-b6e6-9a2f-b8cdd5da87b3, 'name': SearchDatastore_Task, 'duration_secs': 0.013382} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.756481] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.756736] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.756980] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.757155] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.757680] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.757680] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a8f9c78-fdfe-4c2b-b471-9a28a392ea82 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.766500] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.766626] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 830.767428] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93c5d69a-f688-4796-ac45-66b44eac11de {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.775893] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 830.775893] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5203969b-aaac-4f4c-e9ad-a64905b1828c" [ 830.775893] env[63293]: _type = "Task" [ 830.775893] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.784373] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5203969b-aaac-4f4c-e9ad-a64905b1828c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.867623] env[63293]: DEBUG nova.compute.manager [req-0c2a9f5c-156f-470b-84fe-9060de3225ce req-bc3f7cd4-9305-4ad1-9ae6-a95e16f2b77a service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Received event network-changed-aa552093-dd1a-4dd1-aadf-9c92b766b4ea {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.867853] env[63293]: DEBUG nova.compute.manager [req-0c2a9f5c-156f-470b-84fe-9060de3225ce req-bc3f7cd4-9305-4ad1-9ae6-a95e16f2b77a service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Refreshing instance network info cache due to event network-changed-aa552093-dd1a-4dd1-aadf-9c92b766b4ea. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 830.868379] env[63293]: DEBUG oslo_concurrency.lockutils [req-0c2a9f5c-156f-470b-84fe-9060de3225ce req-bc3f7cd4-9305-4ad1-9ae6-a95e16f2b77a service nova] Acquiring lock "refresh_cache-b8637881-9fb6-4c5d-848d-7b2d38f8e970" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.868691] env[63293]: DEBUG oslo_concurrency.lockutils [req-0c2a9f5c-156f-470b-84fe-9060de3225ce req-bc3f7cd4-9305-4ad1-9ae6-a95e16f2b77a service nova] Acquired lock "refresh_cache-b8637881-9fb6-4c5d-848d-7b2d38f8e970" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.871590] env[63293]: DEBUG nova.network.neutron [req-0c2a9f5c-156f-470b-84fe-9060de3225ce req-bc3f7cd4-9305-4ad1-9ae6-a95e16f2b77a service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Refreshing network info cache for port aa552093-dd1a-4dd1-aadf-9c92b766b4ea {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 831.071352] env[63293]: DEBUG nova.scheduler.client.report [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.167754] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327597, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.179051] env[63293]: DEBUG nova.compute.manager [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.181218] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.190173] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Releasing lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.190500] env[63293]: DEBUG nova.compute.manager [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Instance network_info: |[{"id": "52a5b959-25ef-48ed-ad04-82f85a131929", "address": "fa:16:3e:56:4e:4a", "network": {"id": "b6024f78-d9d9-4fd8-a71e-9f996ce4a30f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-959509061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76803a169e7d424ab6aefe0d772b7827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52a5b959-25", "ovs_interfaceid": "52a5b959-25ef-48ed-ad04-82f85a131929", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 831.190920] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:4e:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f856fca-9fb5-41ea-a057-ac4193bd323d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52a5b959-25ef-48ed-ad04-82f85a131929', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.203386] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Creating folder: Project (76803a169e7d424ab6aefe0d772b7827). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 831.205715] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c36b4b35-38e7-45eb-a8e2-c4d50da253ae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.216902] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Created folder: Project (76803a169e7d424ab6aefe0d772b7827) in parent group-v283678. [ 831.217153] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Creating folder: Instances. Parent ref: group-v283733. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 831.217412] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c59fe424-f9e2-4ae3-857f-e5ac0ecb58ae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.225264] env[63293]: DEBUG nova.virt.hardware [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.225264] env[63293]: DEBUG nova.virt.hardware [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.225264] env[63293]: DEBUG nova.virt.hardware [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.225548] env[63293]: DEBUG nova.virt.hardware [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.225548] env[63293]: DEBUG nova.virt.hardware [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.225548] env[63293]: DEBUG nova.virt.hardware [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.225548] env[63293]: DEBUG nova.virt.hardware [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.225548] env[63293]: DEBUG nova.virt.hardware [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.225702] env[63293]: DEBUG nova.virt.hardware [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.225702] env[63293]: DEBUG nova.virt.hardware [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.225702] env[63293]: DEBUG nova.virt.hardware [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.226816] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd36847-3d25-484c-ac4f-39f83fbabe55 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.230546] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Created folder: Instances in parent group-v283733. [ 831.230779] env[63293]: DEBUG oslo.service.loopingcall [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.231338] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 831.231567] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2aa77b4c-c48d-4347-97b7-787f9319b607 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.250261] env[63293]: DEBUG nova.virt.hardware [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.250594] env[63293]: DEBUG nova.virt.hardware [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.250812] env[63293]: DEBUG nova.virt.hardware [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.251028] env[63293]: DEBUG nova.virt.hardware [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.251288] env[63293]: DEBUG nova.virt.hardware [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.251497] env[63293]: DEBUG nova.virt.hardware [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.251953] env[63293]: DEBUG nova.virt.hardware [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.252186] env[63293]: DEBUG nova.virt.hardware [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.252374] env[63293]: DEBUG nova.virt.hardware [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.252545] env[63293]: DEBUG nova.virt.hardware [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.252724] env[63293]: DEBUG nova.virt.hardware [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.256635] env[63293]: DEBUG nova.compute.manager [req-d883deeb-2184-4484-b51e-0dfec08dccf3 req-81f133e7-6630-4fb3-83e7-e52f0d5deba3 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Received event network-changed-52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.256817] env[63293]: DEBUG nova.compute.manager [req-d883deeb-2184-4484-b51e-0dfec08dccf3 req-81f133e7-6630-4fb3-83e7-e52f0d5deba3 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Refreshing instance network info cache due to event network-changed-52a5b959-25ef-48ed-ad04-82f85a131929. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.257101] env[63293]: DEBUG oslo_concurrency.lockutils [req-d883deeb-2184-4484-b51e-0dfec08dccf3 req-81f133e7-6630-4fb3-83e7-e52f0d5deba3 service nova] Acquiring lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.257307] env[63293]: DEBUG oslo_concurrency.lockutils [req-d883deeb-2184-4484-b51e-0dfec08dccf3 req-81f133e7-6630-4fb3-83e7-e52f0d5deba3 service nova] Acquired lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.257546] env[63293]: DEBUG nova.network.neutron [req-d883deeb-2184-4484-b51e-0dfec08dccf3 req-81f133e7-6630-4fb3-83e7-e52f0d5deba3 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Refreshing network info cache for port 52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 831.267038] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9175bfdf-6abd-464e-92d4-293b92a0827a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.270534] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8866b8ab-0c85-4a28-a82a-49baa76d20a2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.276852] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.276852] env[63293]: value = "task-1327600" [ 831.276852] env[63293]: _type = "Task" [ 831.276852] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.295705] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:ac:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aec0089a-ff85-4bef-bad8-c84de39af71a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01c58487-3d79-41c6-be70-8e4c5f2efe50', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.303146] env[63293]: DEBUG oslo.service.loopingcall [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.304584] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 831.305902] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d25a7be-46da-4a68-9780-6a354bdec82d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.319319] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5fbd8feb-c6bb-417f-96ea-8e9a95c036eb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.332198] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5203969b-aaac-4f4c-e9ad-a64905b1828c, 'name': SearchDatastore_Task, 'duration_secs': 0.014073} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.332198] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327600, 'name': CreateVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.333830] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb198580-051a-4527-9188-0508ad035c49 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.346493] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.346493] env[63293]: value = "task-1327601" [ 831.346493] env[63293]: _type = "Task" [ 831.346493] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.351689] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 831.351689] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]521a299e-55e4-f93e-9690-243ccde8e349" [ 831.351689] env[63293]: _type = "Task" [ 831.351689] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.358776] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327601, 'name': CreateVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.366464] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]521a299e-55e4-f93e-9690-243ccde8e349, 'name': SearchDatastore_Task} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.366858] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.367244] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] ad585ebb-2072-45df-b645-94c9fa93576b/ad585ebb-2072-45df-b645-94c9fa93576b.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 831.367593] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-25e6089d-d466-431e-b7fa-805da533fe72 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.373699] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 831.373699] env[63293]: value = "task-1327602" [ 831.373699] env[63293]: _type = "Task" [ 831.373699] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.384963] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327602, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.577285] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.577285] env[63293]: DEBUG nova.compute.manager [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 831.580059] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.656s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.581760] env[63293]: INFO nova.compute.claims [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.670198] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327597, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.678984] env[63293]: DEBUG nova.network.neutron [req-0c2a9f5c-156f-470b-84fe-9060de3225ce req-bc3f7cd4-9305-4ad1-9ae6-a95e16f2b77a service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Updated VIF entry in instance network info cache for port aa552093-dd1a-4dd1-aadf-9c92b766b4ea. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 831.679431] env[63293]: DEBUG nova.network.neutron [req-0c2a9f5c-156f-470b-84fe-9060de3225ce req-bc3f7cd4-9305-4ad1-9ae6-a95e16f2b77a service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Updating instance_info_cache with network_info: [{"id": "aa552093-dd1a-4dd1-aadf-9c92b766b4ea", "address": "fa:16:3e:13:f8:5d", "network": {"id": "0ec4e867-7095-4c29-9202-669ad2701e74", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-378057015-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fd63e6892c44f159195897ad5eacf13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa552093-dd", "ovs_interfaceid": "aa552093-dd1a-4dd1-aadf-9c92b766b4ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.788017] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327600, 'name': CreateVM_Task, 'duration_secs': 0.408526} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.788312] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 831.788990] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.789175] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.789557] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.789899] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9abcc95b-a1b5-433c-a250-153510521e84 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.795304] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 831.795304] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a050fe-3434-ae76-275a-6f11877ceff4" [ 831.795304] env[63293]: _type = "Task" [ 831.795304] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.809201] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a050fe-3434-ae76-275a-6f11877ceff4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.858586] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327601, 'name': CreateVM_Task, 'duration_secs': 0.386779} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.861375] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 831.861901] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.884591] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327602, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.032112] env[63293]: DEBUG nova.network.neutron [req-d883deeb-2184-4484-b51e-0dfec08dccf3 req-81f133e7-6630-4fb3-83e7-e52f0d5deba3 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updated VIF entry in instance network info cache for port 52a5b959-25ef-48ed-ad04-82f85a131929. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 832.032617] env[63293]: DEBUG nova.network.neutron [req-d883deeb-2184-4484-b51e-0dfec08dccf3 req-81f133e7-6630-4fb3-83e7-e52f0d5deba3 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updating instance_info_cache with network_info: [{"id": "52a5b959-25ef-48ed-ad04-82f85a131929", "address": "fa:16:3e:56:4e:4a", "network": {"id": "b6024f78-d9d9-4fd8-a71e-9f996ce4a30f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-959509061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76803a169e7d424ab6aefe0d772b7827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52a5b959-25", "ovs_interfaceid": "52a5b959-25ef-48ed-ad04-82f85a131929", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.100653] env[63293]: DEBUG nova.compute.utils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.109588] env[63293]: DEBUG nova.compute.manager [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 832.109588] env[63293]: DEBUG nova.network.neutron [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 832.149855] env[63293]: DEBUG nova.policy [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7bf93621dae48ceafcf2a16e95ec84e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '92db4287612b4b5f912afcd4a8d187ed', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 832.168328] env[63293]: DEBUG oslo_vmware.api [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327597, 'name': PowerOnVM_Task, 'duration_secs': 1.503283} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.168628] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 832.168812] env[63293]: INFO nova.compute.manager [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Took 10.53 seconds to spawn the instance on the hypervisor. [ 832.169024] env[63293]: DEBUG nova.compute.manager [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.169863] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f158b737-93ef-451e-8caf-30e239bc9036 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.182575] env[63293]: DEBUG oslo_concurrency.lockutils [req-0c2a9f5c-156f-470b-84fe-9060de3225ce req-bc3f7cd4-9305-4ad1-9ae6-a95e16f2b77a service nova] Releasing lock "refresh_cache-b8637881-9fb6-4c5d-848d-7b2d38f8e970" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.236700] env[63293]: DEBUG nova.compute.manager [req-7fef8b3a-451d-410a-828a-254e7bee4f92 req-87ef4566-de52-44a7-96de-0ddc4c6ab75c service nova] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Received event network-vif-plugged-1582d90f-11d5-4ec4-8173-e1fcb11d3bc8 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.236929] env[63293]: DEBUG oslo_concurrency.lockutils [req-7fef8b3a-451d-410a-828a-254e7bee4f92 req-87ef4566-de52-44a7-96de-0ddc4c6ab75c service nova] Acquiring lock "1e01a75d-122d-4122-9fed-4164d64d4ee8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.237268] env[63293]: DEBUG oslo_concurrency.lockutils [req-7fef8b3a-451d-410a-828a-254e7bee4f92 req-87ef4566-de52-44a7-96de-0ddc4c6ab75c service nova] Lock "1e01a75d-122d-4122-9fed-4164d64d4ee8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.237457] env[63293]: DEBUG oslo_concurrency.lockutils [req-7fef8b3a-451d-410a-828a-254e7bee4f92 req-87ef4566-de52-44a7-96de-0ddc4c6ab75c service nova] Lock "1e01a75d-122d-4122-9fed-4164d64d4ee8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.237644] env[63293]: DEBUG nova.compute.manager [req-7fef8b3a-451d-410a-828a-254e7bee4f92 req-87ef4566-de52-44a7-96de-0ddc4c6ab75c service nova] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] No waiting events found dispatching network-vif-plugged-1582d90f-11d5-4ec4-8173-e1fcb11d3bc8 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 832.237817] env[63293]: WARNING nova.compute.manager [req-7fef8b3a-451d-410a-828a-254e7bee4f92 req-87ef4566-de52-44a7-96de-0ddc4c6ab75c service nova] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Received unexpected event network-vif-plugged-1582d90f-11d5-4ec4-8173-e1fcb11d3bc8 for instance with vm_state building and task_state spawning. [ 832.306750] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a050fe-3434-ae76-275a-6f11877ceff4, 'name': SearchDatastore_Task, 'duration_secs': 0.021782} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.307094] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.307381] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.307628] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.307777] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.307962] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.308279] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.308597] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 832.309063] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b725911a-fad9-42de-b9a7-b606bd15dde9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.311014] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40c96e6d-450b-4ac2-beb7-c9f23a77f6b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.315949] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 832.315949] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522e5e2a-83fc-38a3-4503-563f17bb6beb" [ 832.315949] env[63293]: _type = "Task" [ 832.315949] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.320927] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.321386] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 832.322610] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfe05855-eb66-4df2-b4c0-a3d1fcff627a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.328011] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522e5e2a-83fc-38a3-4503-563f17bb6beb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.331392] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 832.331392] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52aa6dd4-354f-8c4c-4fa0-3371cb194cda" [ 832.331392] env[63293]: _type = "Task" [ 832.331392] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.339811] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52aa6dd4-354f-8c4c-4fa0-3371cb194cda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.384666] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327602, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.651367} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.384992] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] ad585ebb-2072-45df-b645-94c9fa93576b/ad585ebb-2072-45df-b645-94c9fa93576b.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 832.385270] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.385625] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-49578017-9760-4f05-9062-890b2da359ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.393379] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 832.393379] env[63293]: value = "task-1327603" [ 832.393379] env[63293]: _type = "Task" [ 832.393379] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.403634] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327603, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.535915] env[63293]: DEBUG oslo_concurrency.lockutils [req-d883deeb-2184-4484-b51e-0dfec08dccf3 req-81f133e7-6630-4fb3-83e7-e52f0d5deba3 service nova] Releasing lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.586127] env[63293]: DEBUG nova.network.neutron [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Successfully created port: f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.609529] env[63293]: DEBUG nova.compute.manager [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 832.688139] env[63293]: INFO nova.compute.manager [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Took 32.82 seconds to build instance. [ 832.832280] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522e5e2a-83fc-38a3-4503-563f17bb6beb, 'name': SearchDatastore_Task, 'duration_secs': 0.012286} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.836808] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.837609] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.838021] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.847023] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52aa6dd4-354f-8c4c-4fa0-3371cb194cda, 'name': SearchDatastore_Task, 'duration_secs': 0.011829} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.847023] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7874febc-ec39-4e19-a26c-1c12d7df521b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.855027] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 832.855027] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523f916e-5326-d0a2-f454-cd3d9e771f84" [ 832.855027] env[63293]: _type = "Task" [ 832.855027] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.870531] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523f916e-5326-d0a2-f454-cd3d9e771f84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.871030] env[63293]: DEBUG nova.network.neutron [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Successfully updated port: 1582d90f-11d5-4ec4-8173-e1fcb11d3bc8 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.909606] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327603, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068061} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.912495] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.913627] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce7fd76-9051-4682-85d7-a04ef708c12f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.936271] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] ad585ebb-2072-45df-b645-94c9fa93576b/ad585ebb-2072-45df-b645-94c9fa93576b.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.939188] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e93d530e-6147-4908-8205-82cb9c4040ad {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.962116] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 832.962116] env[63293]: value = "task-1327604" [ 832.962116] env[63293]: _type = "Task" [ 832.962116] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.970126] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327604, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.012204] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb21150f-e5d8-4b74-abe8-a76b24156687 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.020078] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a619baa0-2297-4ab4-a617-5e4b2ea3abf9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.050710] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85edc4a-42b2-434b-8093-c81ce03e1a2d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.058624] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c72f226-e082-4629-86c6-2f31b07c2b7c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.072648] env[63293]: DEBUG nova.compute.provider_tree [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.190819] env[63293]: DEBUG oslo_concurrency.lockutils [None req-176342fa-d46a-423b-8bc0-07127b56b546 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.703s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.288464] env[63293]: DEBUG nova.compute.manager [req-43a601db-aeec-4dcb-a535-166500cd45f1 req-6e9707db-e7cc-4ff6-a10c-eaf0547c8a3c service nova] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Received event network-changed-1582d90f-11d5-4ec4-8173-e1fcb11d3bc8 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.288680] env[63293]: DEBUG nova.compute.manager [req-43a601db-aeec-4dcb-a535-166500cd45f1 req-6e9707db-e7cc-4ff6-a10c-eaf0547c8a3c service nova] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Refreshing instance network info cache due to event network-changed-1582d90f-11d5-4ec4-8173-e1fcb11d3bc8. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.289239] env[63293]: DEBUG oslo_concurrency.lockutils [req-43a601db-aeec-4dcb-a535-166500cd45f1 req-6e9707db-e7cc-4ff6-a10c-eaf0547c8a3c service nova] Acquiring lock "refresh_cache-1e01a75d-122d-4122-9fed-4164d64d4ee8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.289239] env[63293]: DEBUG oslo_concurrency.lockutils [req-43a601db-aeec-4dcb-a535-166500cd45f1 req-6e9707db-e7cc-4ff6-a10c-eaf0547c8a3c service nova] Acquired lock "refresh_cache-1e01a75d-122d-4122-9fed-4164d64d4ee8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.289396] env[63293]: DEBUG nova.network.neutron [req-43a601db-aeec-4dcb-a535-166500cd45f1 req-6e9707db-e7cc-4ff6-a10c-eaf0547c8a3c service nova] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Refreshing network info cache for port 1582d90f-11d5-4ec4-8173-e1fcb11d3bc8 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.365467] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523f916e-5326-d0a2-f454-cd3d9e771f84, 'name': SearchDatastore_Task, 'duration_secs': 0.019205} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.365597] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.365892] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 831d7d9b-364f-4c29-bda5-e6c38291a973/831d7d9b-364f-4c29-bda5-e6c38291a973.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 833.366614] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.366614] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.366614] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bbf2035a-e458-4c96-b419-bd85223c1aa4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.369501] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed6bec5f-0ca4-449a-9dee-6cee2115cd57 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.373273] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquiring lock "refresh_cache-1e01a75d-122d-4122-9fed-4164d64d4ee8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.379348] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 833.379348] env[63293]: value = "task-1327605" [ 833.379348] env[63293]: _type = "Task" [ 833.379348] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.382142] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.382329] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 833.388944] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88fc4271-0131-453e-8877-e90781d2e955 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.400729] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.402878] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 833.402878] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]521aad60-fc2b-90cf-2b60-0410d4ab44ce" [ 833.402878] env[63293]: _type = "Task" [ 833.402878] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.410214] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]521aad60-fc2b-90cf-2b60-0410d4ab44ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.420762] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521315ef-5328-4c62-2280-221bd775fa1e/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 833.421710] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6eef51-375c-4fe2-813d-d0e58789b2ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.428383] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521315ef-5328-4c62-2280-221bd775fa1e/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 833.428615] env[63293]: ERROR oslo_vmware.rw_handles [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521315ef-5328-4c62-2280-221bd775fa1e/disk-0.vmdk due to incomplete transfer. [ 833.428863] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9ca57900-7d38-4424-b0ea-a682a945e826 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.437057] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521315ef-5328-4c62-2280-221bd775fa1e/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 833.437284] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Uploaded image 54404016-2794-4f2d-a4d4-8ec6658c7f45 to the Glance image server {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 833.440431] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Destroying the VM {{(pid=63293) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 833.440431] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-671f2afd-8daa-4c3b-9db1-4291771631f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.448147] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 833.448147] env[63293]: value = "task-1327606" [ 833.448147] env[63293]: _type = "Task" [ 833.448147] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.455884] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327606, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.472671] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327604, 'name': ReconfigVM_Task, 'duration_secs': 0.426586} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.472964] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Reconfigured VM instance instance-0000003f to attach disk [datastore2] ad585ebb-2072-45df-b645-94c9fa93576b/ad585ebb-2072-45df-b645-94c9fa93576b.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.473840] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce482a93-702a-4843-9996-473257313cf2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.483051] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 833.483051] env[63293]: value = "task-1327607" [ 833.483051] env[63293]: _type = "Task" [ 833.483051] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.490956] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327607, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.576355] env[63293]: DEBUG nova.scheduler.client.report [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.619924] env[63293]: DEBUG nova.compute.manager [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 833.648405] env[63293]: DEBUG nova.virt.hardware [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.648696] env[63293]: DEBUG nova.virt.hardware [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.648864] env[63293]: DEBUG nova.virt.hardware [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.649092] env[63293]: DEBUG nova.virt.hardware [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.649270] env[63293]: DEBUG nova.virt.hardware [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.649409] env[63293]: DEBUG nova.virt.hardware [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.649618] env[63293]: DEBUG nova.virt.hardware [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.649794] env[63293]: DEBUG nova.virt.hardware [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.650836] env[63293]: DEBUG nova.virt.hardware [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.650836] env[63293]: DEBUG nova.virt.hardware [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.650836] env[63293]: DEBUG nova.virt.hardware [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.651652] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe0aa5d-fa0c-47ad-8de5-8c9c8b4563eb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.660309] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c77717-ce50-4ff1-bea2-1adb8225a3a8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.696590] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 833.851939] env[63293]: DEBUG nova.network.neutron [req-43a601db-aeec-4dcb-a535-166500cd45f1 req-6e9707db-e7cc-4ff6-a10c-eaf0547c8a3c service nova] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.895598] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327605, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.917619] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]521aad60-fc2b-90cf-2b60-0410d4ab44ce, 'name': SearchDatastore_Task, 'duration_secs': 0.043985} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.917619] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff2c61a1-83a7-40e4-87a4-7d8cae382686 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.929030] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 833.929030] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5207ac3e-685a-2de4-cc9a-34f40589632f" [ 833.929030] env[63293]: _type = "Task" [ 833.929030] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.935721] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5207ac3e-685a-2de4-cc9a-34f40589632f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.960672] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327606, 'name': Destroy_Task, 'duration_secs': 0.328391} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.961415] env[63293]: DEBUG nova.network.neutron [req-43a601db-aeec-4dcb-a535-166500cd45f1 req-6e9707db-e7cc-4ff6-a10c-eaf0547c8a3c service nova] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.962850] env[63293]: INFO nova.compute.manager [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Rescuing [ 833.963127] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquiring lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.963298] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquired lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.963468] env[63293]: DEBUG nova.network.neutron [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.964570] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Destroyed the VM [ 833.964928] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Deleting Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 833.965500] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f86d0695-6633-4250-874c-53883984c0e1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.975143] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 833.975143] env[63293]: value = "task-1327608" [ 833.975143] env[63293]: _type = "Task" [ 833.975143] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.989401] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327608, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.994838] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327607, 'name': Rename_Task, 'duration_secs': 0.149543} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.995205] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 833.995473] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f253628b-5fbb-40f3-8833-b2231b0dad52 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.005869] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 834.005869] env[63293]: value = "task-1327609" [ 834.005869] env[63293]: _type = "Task" [ 834.005869] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.016648] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327609, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.083152] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.083152] env[63293]: DEBUG nova.compute.manager [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 834.088369] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.531s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.089940] env[63293]: INFO nova.compute.claims [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.223655] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.310637] env[63293]: DEBUG nova.network.neutron [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Successfully updated port: f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.394783] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327605, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.756132} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.395067] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 831d7d9b-364f-4c29-bda5-e6c38291a973/831d7d9b-364f-4c29-bda5-e6c38291a973.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 834.395295] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.395559] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-638cbcbe-de35-47d1-a88c-8e078cb1c27a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.405605] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 834.405605] env[63293]: value = "task-1327610" [ 834.405605] env[63293]: _type = "Task" [ 834.405605] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.416018] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327610, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.437989] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5207ac3e-685a-2de4-cc9a-34f40589632f, 'name': SearchDatastore_Task, 'duration_secs': 0.056768} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.438895] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.438895] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 33d71260-26f1-482c-b93b-5f8e27c545f3/33d71260-26f1-482c-b93b-5f8e27c545f3.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 834.438895] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d31a297b-489e-4ca9-b26f-647e946b2b27 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.443931] env[63293]: DEBUG nova.compute.manager [req-28092ba4-3743-482d-99c8-12a75409de74 req-2bd00cd9-f6ac-4134-b888-eb25d3b56181 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Received event network-vif-plugged-f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.443931] env[63293]: DEBUG oslo_concurrency.lockutils [req-28092ba4-3743-482d-99c8-12a75409de74 req-2bd00cd9-f6ac-4134-b888-eb25d3b56181 service nova] Acquiring lock "4593cabd-1331-4ee1-99f5-46edeac56859-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.444132] env[63293]: DEBUG oslo_concurrency.lockutils [req-28092ba4-3743-482d-99c8-12a75409de74 req-2bd00cd9-f6ac-4134-b888-eb25d3b56181 service nova] Lock "4593cabd-1331-4ee1-99f5-46edeac56859-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.444304] env[63293]: DEBUG oslo_concurrency.lockutils [req-28092ba4-3743-482d-99c8-12a75409de74 req-2bd00cd9-f6ac-4134-b888-eb25d3b56181 service nova] Lock "4593cabd-1331-4ee1-99f5-46edeac56859-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.444536] env[63293]: DEBUG nova.compute.manager [req-28092ba4-3743-482d-99c8-12a75409de74 req-2bd00cd9-f6ac-4134-b888-eb25d3b56181 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] No waiting events found dispatching network-vif-plugged-f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 834.444836] env[63293]: WARNING nova.compute.manager [req-28092ba4-3743-482d-99c8-12a75409de74 req-2bd00cd9-f6ac-4134-b888-eb25d3b56181 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Received unexpected event network-vif-plugged-f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13 for instance with vm_state building and task_state spawning. [ 834.448018] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 834.448018] env[63293]: value = "task-1327611" [ 834.448018] env[63293]: _type = "Task" [ 834.448018] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.456404] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327611, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.465761] env[63293]: DEBUG oslo_concurrency.lockutils [req-43a601db-aeec-4dcb-a535-166500cd45f1 req-6e9707db-e7cc-4ff6-a10c-eaf0547c8a3c service nova] Releasing lock "refresh_cache-1e01a75d-122d-4122-9fed-4164d64d4ee8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.466147] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquired lock "refresh_cache-1e01a75d-122d-4122-9fed-4164d64d4ee8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.466286] env[63293]: DEBUG nova.network.neutron [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.485352] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327608, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.523784] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327609, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.594302] env[63293]: DEBUG nova.compute.utils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 834.601245] env[63293]: DEBUG nova.compute.manager [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 834.601439] env[63293]: DEBUG nova.network.neutron [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 834.656727] env[63293]: DEBUG nova.policy [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e70e83ed8f74a81ab2607ed1a20484e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e5d75a3118c4a378f3a1ac529f5c5dc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 834.791044] env[63293]: DEBUG nova.network.neutron [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updating instance_info_cache with network_info: [{"id": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "address": "fa:16:3e:fc:c1:7d", "network": {"id": "bbf31454-8e46-45d9-b10a-fccda17654c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1875587777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "6fa6cd8dea16412db105faf2ff93a6ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26b86f34-f9", "ovs_interfaceid": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.813740] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquiring lock "refresh_cache-4593cabd-1331-4ee1-99f5-46edeac56859" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.814282] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquired lock "refresh_cache-4593cabd-1331-4ee1-99f5-46edeac56859" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.814282] env[63293]: DEBUG nova.network.neutron [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.918443] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327610, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127178} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.918608] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.919699] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d2640c-d5f3-47cf-a56e-8f6cd3103707 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.947035] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 831d7d9b-364f-4c29-bda5-e6c38291a973/831d7d9b-364f-4c29-bda5-e6c38291a973.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.947501] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d37b09a-7313-4f97-99e5-6317305da0af {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.987403] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 834.987403] env[63293]: value = "task-1327612" [ 834.987403] env[63293]: _type = "Task" [ 834.987403] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.000098] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327611, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.000098] env[63293]: DEBUG oslo_vmware.api [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327608, 'name': RemoveSnapshot_Task, 'duration_secs': 0.944222} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.000326] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Deleted Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 835.004776] env[63293]: INFO nova.compute.manager [None req-2c2dc2b1-76cc-40f2-ab4f-584fd2a207e1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Took 15.44 seconds to snapshot the instance on the hypervisor. [ 835.009481] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.020941] env[63293]: DEBUG oslo_vmware.api [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327609, 'name': PowerOnVM_Task, 'duration_secs': 0.814865} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.022185] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 835.022185] env[63293]: INFO nova.compute.manager [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Took 10.92 seconds to spawn the instance on the hypervisor. [ 835.022185] env[63293]: DEBUG nova.compute.manager [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.022658] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e043ee4-9c89-4118-9799-ac930538a8be {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.038463] env[63293]: DEBUG nova.network.neutron [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.081286] env[63293]: DEBUG nova.network.neutron [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Successfully created port: 20da136a-9449-4b46-be97-b9be3c4c2bfe {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 835.106716] env[63293]: DEBUG nova.compute.manager [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 835.224131] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquiring lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.224131] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.249924] env[63293]: DEBUG nova.network.neutron [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Updating instance_info_cache with network_info: [{"id": "1582d90f-11d5-4ec4-8173-e1fcb11d3bc8", "address": "fa:16:3e:9f:c6:ab", "network": {"id": "7043dd5d-3429-4206-ad79-29343710d8e6", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-634465850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b461ff81505c49dbbd9eb64935492b43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1582d90f-11", "ovs_interfaceid": "1582d90f-11d5-4ec4-8173-e1fcb11d3bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.296568] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Releasing lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.384338] env[63293]: DEBUG nova.network.neutron [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.457491] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327611, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.711659} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.460327] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 33d71260-26f1-482c-b93b-5f8e27c545f3/33d71260-26f1-482c-b93b-5f8e27c545f3.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 835.460464] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.461119] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45511b2b-fcbb-4b71-aa7e-a32ed1a82ff5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.470608] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 835.470608] env[63293]: value = "task-1327613" [ 835.470608] env[63293]: _type = "Task" [ 835.470608] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.478365] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327613, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.487396] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b54d9c0-ff2c-4c94-a804-56175c26eea8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.504447] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.506921] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1411ac63-f053-4c25-a483-f0990217eb57 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.549453] env[63293]: INFO nova.compute.manager [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Took 26.89 seconds to build instance. [ 835.553034] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984c2986-8b9b-45fa-a853-fe3a773fd037 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.564977] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d355d3-8ad2-499b-9270-e9e213f8881d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.581817] env[63293]: DEBUG nova.compute.provider_tree [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.606643] env[63293]: DEBUG nova.network.neutron [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Updating instance_info_cache with network_info: [{"id": "f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13", "address": "fa:16:3e:93:a3:f3", "network": {"id": "22d4ebf1-8d5d-43c5-a2b9-24c55f8b22bb", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-987787356-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92db4287612b4b5f912afcd4a8d187ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7aa6264-122d-4c35-82d0-860e451538ea", "external-id": "nsx-vlan-transportzone-81", "segmentation_id": 81, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8b1091f-58", "ovs_interfaceid": "f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.753141] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Releasing lock "refresh_cache-1e01a75d-122d-4122-9fed-4164d64d4ee8" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.753492] env[63293]: DEBUG nova.compute.manager [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Instance network_info: |[{"id": "1582d90f-11d5-4ec4-8173-e1fcb11d3bc8", "address": "fa:16:3e:9f:c6:ab", "network": {"id": "7043dd5d-3429-4206-ad79-29343710d8e6", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-634465850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b461ff81505c49dbbd9eb64935492b43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1582d90f-11", "ovs_interfaceid": "1582d90f-11d5-4ec4-8173-e1fcb11d3bc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 835.753911] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:c6:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7b4bfde-f109-4f64-adab-e7f06b80685d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1582d90f-11d5-4ec4-8173-e1fcb11d3bc8', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.761799] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Creating folder: Project (b461ff81505c49dbbd9eb64935492b43). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.762108] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1303bef4-5d6b-42a6-9f09-e71c492b5d01 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.773551] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Created folder: Project (b461ff81505c49dbbd9eb64935492b43) in parent group-v283678. [ 835.773818] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Creating folder: Instances. Parent ref: group-v283737. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.774152] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-26ee5a5b-002a-4e2f-a9f3-77a2653892af {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.783545] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Created folder: Instances in parent group-v283737. [ 835.783813] env[63293]: DEBUG oslo.service.loopingcall [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.784046] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.784306] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee96512a-29a4-4b8b-983e-54f6ad75bb42 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.806095] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.806095] env[63293]: value = "task-1327616" [ 835.806095] env[63293]: _type = "Task" [ 835.806095] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.814533] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327616, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.833969] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.834959] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.836904] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.837515] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.838069] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.842611] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 835.844053] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a3a3a048-b3c9-40e5-953a-cd5076ec3a89 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.847655] env[63293]: INFO nova.compute.manager [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Terminating instance [ 835.851590] env[63293]: DEBUG nova.compute.manager [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 835.852021] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 835.856711] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66840b3-a7a7-4fdc-889f-c3c5e028649d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.860862] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 835.860862] env[63293]: value = "task-1327617" [ 835.860862] env[63293]: _type = "Task" [ 835.860862] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.869053] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.869881] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ffac7fe7-9853-4633-a49b-9968997d4573 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.875646] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327617, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.979327] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327613, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069516} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.980650] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.983830] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9bd8dfd-aabf-424c-83eb-305c9ccebb55 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.011190] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 33d71260-26f1-482c-b93b-5f8e27c545f3/33d71260-26f1-482c-b93b-5f8e27c545f3.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.016166] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4b684d0-b34f-4263-9ddd-9d5f55c4f11a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.034085] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 836.034085] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 836.034312] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleting the datastore file [datastore2] cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 836.034776] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-622f333e-adc5-4675-9c2a-bbdbcf9dbd95 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.044178] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327612, 'name': ReconfigVM_Task, 'duration_secs': 1.021615} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.046759] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 831d7d9b-364f-4c29-bda5-e6c38291a973/831d7d9b-364f-4c29-bda5-e6c38291a973.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.048090] env[63293]: DEBUG oslo_vmware.api [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 836.048090] env[63293]: value = "task-1327620" [ 836.048090] env[63293]: _type = "Task" [ 836.048090] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.048363] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 836.048363] env[63293]: value = "task-1327619" [ 836.048363] env[63293]: _type = "Task" [ 836.048363] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.048700] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-75971a9d-3282-4e0b-9458-99424b23ae83 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.060825] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fd177eb4-e40a-454d-af90-271a3e13390c tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.759s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.070292] env[63293]: DEBUG oslo_vmware.api [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327620, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.070795] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 836.070795] env[63293]: value = "task-1327621" [ 836.070795] env[63293]: _type = "Task" [ 836.070795] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.076606] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327619, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.084727] env[63293]: DEBUG nova.scheduler.client.report [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.100024] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327621, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.110716] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Releasing lock "refresh_cache-4593cabd-1331-4ee1-99f5-46edeac56859" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.111094] env[63293]: DEBUG nova.compute.manager [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Instance network_info: |[{"id": "f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13", "address": "fa:16:3e:93:a3:f3", "network": {"id": "22d4ebf1-8d5d-43c5-a2b9-24c55f8b22bb", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-987787356-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92db4287612b4b5f912afcd4a8d187ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7aa6264-122d-4c35-82d0-860e451538ea", "external-id": "nsx-vlan-transportzone-81", "segmentation_id": 81, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8b1091f-58", "ovs_interfaceid": "f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 836.111778] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:a3:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7aa6264-122d-4c35-82d0-860e451538ea', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.119659] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Creating folder: Project (92db4287612b4b5f912afcd4a8d187ed). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.120836] env[63293]: DEBUG nova.compute.manager [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 836.123273] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1deffcc5-eae3-4f2c-8721-c06c80db6a79 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.133680] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Created folder: Project (92db4287612b4b5f912afcd4a8d187ed) in parent group-v283678. [ 836.133917] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Creating folder: Instances. Parent ref: group-v283740. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.134193] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0fd5095-8b86-476d-a41c-ea53a62e437b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.145556] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Created folder: Instances in parent group-v283740. [ 836.145894] env[63293]: DEBUG oslo.service.loopingcall [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.146153] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 836.146395] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-43687188-e204-40c6-b14d-5b217e3b404b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.165207] env[63293]: DEBUG nova.virt.hardware [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.165414] env[63293]: DEBUG nova.virt.hardware [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.165547] env[63293]: DEBUG nova.virt.hardware [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.165797] env[63293]: DEBUG nova.virt.hardware [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.166020] env[63293]: DEBUG nova.virt.hardware [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.166230] env[63293]: DEBUG nova.virt.hardware [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.166478] env[63293]: DEBUG nova.virt.hardware [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.166691] env[63293]: DEBUG nova.virt.hardware [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.166933] env[63293]: DEBUG nova.virt.hardware [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.167194] env[63293]: DEBUG nova.virt.hardware [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.167437] env[63293]: DEBUG nova.virt.hardware [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.168333] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86c526a-71f6-470f-bd70-3bc78bfd0259 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.180596] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a3a3d6-55f9-4c40-991a-1cdcf317e828 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.185866] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.185866] env[63293]: value = "task-1327624" [ 836.185866] env[63293]: _type = "Task" [ 836.185866] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.199794] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327624, 'name': CreateVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.316026] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327616, 'name': CreateVM_Task, 'duration_secs': 0.376773} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.316341] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 836.317029] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.317224] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.317541] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 836.317798] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-881c966d-8c21-45a4-ae7c-e830dc553af5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.323017] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for the task: (returnval){ [ 836.323017] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520c79f4-95e3-909c-03b6-cd34f4ce6add" [ 836.323017] env[63293]: _type = "Task" [ 836.323017] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.331172] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520c79f4-95e3-909c-03b6-cd34f4ce6add, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.372008] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327617, 'name': PowerOffVM_Task, 'duration_secs': 0.331835} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.372153] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 836.372865] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d03a6b7-4bc9-4a37-8314-cde19616015a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.399563] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3d07be-bcba-4850-a690-17759b3597d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.426958] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.427312] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-595f278b-f330-40e7-83c2-cfea83f45271 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.435230] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 836.435230] env[63293]: value = "task-1327625" [ 836.435230] env[63293]: _type = "Task" [ 836.435230] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.444602] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] VM already powered off {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 836.444916] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.445433] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.446093] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.446093] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.446327] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d528257-8d12-406f-b26e-74495bae91bb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.454799] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.454996] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.455737] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-338588b7-bf01-4bad-a0d3-4f9dcb857ade {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.461625] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 836.461625] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529e0363-36a3-4c8a-20e4-d1ec02177fb0" [ 836.461625] env[63293]: _type = "Task" [ 836.461625] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.469474] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529e0363-36a3-4c8a-20e4-d1ec02177fb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.513776] env[63293]: DEBUG nova.compute.manager [req-8d872f42-1d7c-4e3a-b096-1b73527dee8e req-b20bd3fc-b63b-46cb-b7f6-d3f0c33d48f7 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Received event network-changed-f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.514178] env[63293]: DEBUG nova.compute.manager [req-8d872f42-1d7c-4e3a-b096-1b73527dee8e req-b20bd3fc-b63b-46cb-b7f6-d3f0c33d48f7 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Refreshing instance network info cache due to event network-changed-f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 836.514514] env[63293]: DEBUG oslo_concurrency.lockutils [req-8d872f42-1d7c-4e3a-b096-1b73527dee8e req-b20bd3fc-b63b-46cb-b7f6-d3f0c33d48f7 service nova] Acquiring lock "refresh_cache-4593cabd-1331-4ee1-99f5-46edeac56859" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.514795] env[63293]: DEBUG oslo_concurrency.lockutils [req-8d872f42-1d7c-4e3a-b096-1b73527dee8e req-b20bd3fc-b63b-46cb-b7f6-d3f0c33d48f7 service nova] Acquired lock "refresh_cache-4593cabd-1331-4ee1-99f5-46edeac56859" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.515025] env[63293]: DEBUG nova.network.neutron [req-8d872f42-1d7c-4e3a-b096-1b73527dee8e req-b20bd3fc-b63b-46cb-b7f6-d3f0c33d48f7 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Refreshing network info cache for port f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 836.533111] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.533438] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.564455] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 836.566461] env[63293]: DEBUG oslo_vmware.api [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327620, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.431408} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.570398] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.571013] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.571013] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.571013] env[63293]: INFO nova.compute.manager [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Took 0.72 seconds to destroy the instance on the hypervisor. [ 836.571332] env[63293]: DEBUG oslo.service.loopingcall [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.571766] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327619, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.571977] env[63293]: DEBUG nova.compute.manager [-] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.572094] env[63293]: DEBUG nova.network.neutron [-] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 836.586941] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327621, 'name': Rename_Task, 'duration_secs': 0.243621} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.587581] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.587900] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-879a7278-4a5f-4855-9241-e34edca83d4f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.591213] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.505s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.591828] env[63293]: DEBUG nova.compute.manager [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.598317] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.310s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.600084] env[63293]: INFO nova.compute.claims [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.603459] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 836.603459] env[63293]: value = "task-1327626" [ 836.603459] env[63293]: _type = "Task" [ 836.603459] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.613752] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327626, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.693882] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327624, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.809202] env[63293]: DEBUG nova.compute.manager [req-dd2a829e-53c2-4b54-8d15-c676775dfd1c req-e31dcb8c-d13b-4740-acd9-2284d3e4e869 service nova] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Received event network-vif-plugged-20da136a-9449-4b46-be97-b9be3c4c2bfe {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.809202] env[63293]: DEBUG oslo_concurrency.lockutils [req-dd2a829e-53c2-4b54-8d15-c676775dfd1c req-e31dcb8c-d13b-4740-acd9-2284d3e4e869 service nova] Acquiring lock "93080446-fd45-4281-bf6e-35d1d964f41e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.809202] env[63293]: DEBUG oslo_concurrency.lockutils [req-dd2a829e-53c2-4b54-8d15-c676775dfd1c req-e31dcb8c-d13b-4740-acd9-2284d3e4e869 service nova] Lock "93080446-fd45-4281-bf6e-35d1d964f41e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.809202] env[63293]: DEBUG oslo_concurrency.lockutils [req-dd2a829e-53c2-4b54-8d15-c676775dfd1c req-e31dcb8c-d13b-4740-acd9-2284d3e4e869 service nova] Lock "93080446-fd45-4281-bf6e-35d1d964f41e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.809202] env[63293]: DEBUG nova.compute.manager [req-dd2a829e-53c2-4b54-8d15-c676775dfd1c req-e31dcb8c-d13b-4740-acd9-2284d3e4e869 service nova] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] No waiting events found dispatching network-vif-plugged-20da136a-9449-4b46-be97-b9be3c4c2bfe {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 836.809366] env[63293]: WARNING nova.compute.manager [req-dd2a829e-53c2-4b54-8d15-c676775dfd1c req-e31dcb8c-d13b-4740-acd9-2284d3e4e869 service nova] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Received unexpected event network-vif-plugged-20da136a-9449-4b46-be97-b9be3c4c2bfe for instance with vm_state building and task_state spawning. [ 836.834321] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520c79f4-95e3-909c-03b6-cd34f4ce6add, 'name': SearchDatastore_Task, 'duration_secs': 0.041463} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.834644] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.834878] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.835218] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.835483] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.835524] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.835756] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-024a2ce6-5852-4207-9df7-d18a34037274 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.840905] env[63293]: DEBUG nova.network.neutron [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Successfully updated port: 20da136a-9449-4b46-be97-b9be3c4c2bfe {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.846374] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.846374] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.846602] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfa18bb1-5163-49cc-aad0-4d3003bfd57a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.854578] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for the task: (returnval){ [ 836.854578] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f05591-7433-8209-9cdb-9d4dbf325332" [ 836.854578] env[63293]: _type = "Task" [ 836.854578] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.863463] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f05591-7433-8209-9cdb-9d4dbf325332, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.974216] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529e0363-36a3-4c8a-20e4-d1ec02177fb0, 'name': SearchDatastore_Task, 'duration_secs': 0.01604} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.978958] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2504a03c-7d57-48d0-85c8-226ce6261401 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.983350] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 836.983350] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5239d517-9295-910e-f3b7-12bc73e54860" [ 836.983350] env[63293]: _type = "Task" [ 836.983350] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.991961] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5239d517-9295-910e-f3b7-12bc73e54860, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.068977] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327619, 'name': ReconfigVM_Task, 'duration_secs': 0.594642} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.069307] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 33d71260-26f1-482c-b93b-5f8e27c545f3/33d71260-26f1-482c-b93b-5f8e27c545f3.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.069958] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a558007-5861-46d9-8553-d40f3d02dea2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.076807] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 837.076807] env[63293]: value = "task-1327627" [ 837.076807] env[63293]: _type = "Task" [ 837.076807] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.089072] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327627, 'name': Rename_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.094527] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.106128] env[63293]: DEBUG nova.compute.utils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.110982] env[63293]: DEBUG nova.compute.manager [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 837.111714] env[63293]: DEBUG nova.network.neutron [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 837.123068] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327626, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.194087] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327624, 'name': CreateVM_Task, 'duration_secs': 0.531311} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.197623] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.197867] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.198067] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.198423] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.199067] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dad12bae-1e98-4055-a727-4f6f0617d76e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.203618] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for the task: (returnval){ [ 837.203618] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5298ce65-db47-b0de-27f3-fd471dbdc3f9" [ 837.203618] env[63293]: _type = "Task" [ 837.203618] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.213188] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5298ce65-db47-b0de-27f3-fd471dbdc3f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.227511] env[63293]: DEBUG nova.policy [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b94ec4d2961a43ca9a331a6051fa774d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7744835de0bc4e30b1b29ee5a439cd0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.311523] env[63293]: DEBUG nova.network.neutron [req-8d872f42-1d7c-4e3a-b096-1b73527dee8e req-b20bd3fc-b63b-46cb-b7f6-d3f0c33d48f7 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Updated VIF entry in instance network info cache for port f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 837.311901] env[63293]: DEBUG nova.network.neutron [req-8d872f42-1d7c-4e3a-b096-1b73527dee8e req-b20bd3fc-b63b-46cb-b7f6-d3f0c33d48f7 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Updating instance_info_cache with network_info: [{"id": "f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13", "address": "fa:16:3e:93:a3:f3", "network": {"id": "22d4ebf1-8d5d-43c5-a2b9-24c55f8b22bb", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-987787356-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "92db4287612b4b5f912afcd4a8d187ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7aa6264-122d-4c35-82d0-860e451538ea", "external-id": "nsx-vlan-transportzone-81", "segmentation_id": 81, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8b1091f-58", "ovs_interfaceid": "f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.347155] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquiring lock "refresh_cache-93080446-fd45-4281-bf6e-35d1d964f41e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.347600] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquired lock "refresh_cache-93080446-fd45-4281-bf6e-35d1d964f41e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.347600] env[63293]: DEBUG nova.network.neutron [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 837.371828] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f05591-7433-8209-9cdb-9d4dbf325332, 'name': SearchDatastore_Task, 'duration_secs': 0.011835} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.372734] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-440eed5c-01f7-4ec8-8cb3-a6a420afc0c6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.378857] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for the task: (returnval){ [ 837.378857] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525faf96-f28a-bdd4-e4d7-121d0b0c53b0" [ 837.378857] env[63293]: _type = "Task" [ 837.378857] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.387443] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525faf96-f28a-bdd4-e4d7-121d0b0c53b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.495677] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5239d517-9295-910e-f3b7-12bc73e54860, 'name': SearchDatastore_Task, 'duration_secs': 0.026858} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.495677] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.495677] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 48603812-c3cc-4dae-b8ba-99b9ac5f7969/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk. {{(pid=63293) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 837.495677] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3316c179-e922-4f91-bd61-1c1493974c16 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.502685] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 837.502685] env[63293]: value = "task-1327628" [ 837.502685] env[63293]: _type = "Task" [ 837.502685] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.512886] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327628, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.589478] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327627, 'name': Rename_Task, 'duration_secs': 0.176037} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.589911] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.590237] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18bc012f-3c5b-4409-a0bf-c5317b61cf60 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.598100] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 837.598100] env[63293]: value = "task-1327629" [ 837.598100] env[63293]: _type = "Task" [ 837.598100] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.610093] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327629, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.619828] env[63293]: DEBUG nova.compute.manager [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.630806] env[63293]: DEBUG oslo_vmware.api [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327626, 'name': PowerOnVM_Task, 'duration_secs': 0.749261} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.631274] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 837.631530] env[63293]: INFO nova.compute.manager [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Took 8.99 seconds to spawn the instance on the hypervisor. [ 837.631767] env[63293]: DEBUG nova.compute.manager [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.632733] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ca133a-92e6-4a99-9209-404445055004 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.718221] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5298ce65-db47-b0de-27f3-fd471dbdc3f9, 'name': SearchDatastore_Task, 'duration_secs': 0.01123} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.722062] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.722702] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.722974] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.725797] env[63293]: DEBUG nova.network.neutron [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Successfully created port: 8b8079ba-2afe-42c9-9536-3b8951b51e00 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.818823] env[63293]: DEBUG oslo_concurrency.lockutils [req-8d872f42-1d7c-4e3a-b096-1b73527dee8e req-b20bd3fc-b63b-46cb-b7f6-d3f0c33d48f7 service nova] Releasing lock "refresh_cache-4593cabd-1331-4ee1-99f5-46edeac56859" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.819923] env[63293]: DEBUG nova.network.neutron [-] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.892284] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525faf96-f28a-bdd4-e4d7-121d0b0c53b0, 'name': SearchDatastore_Task, 'duration_secs': 0.013628} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.892284] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.892284] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 1e01a75d-122d-4122-9fed-4164d64d4ee8/1e01a75d-122d-4122-9fed-4164d64d4ee8.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 837.892284] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.892502] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.892502] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a3c13641-2467-4f0c-953d-21ed62b25bdf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.894300] env[63293]: DEBUG nova.network.neutron [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.896467] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1c9fbe0-d38b-4ccd-80db-a37a608f5c57 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.907525] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for the task: (returnval){ [ 837.907525] env[63293]: value = "task-1327630" [ 837.907525] env[63293]: _type = "Task" [ 837.907525] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.912142] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.912408] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 837.913681] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89d63596-d4d8-48d5-b4fb-350f2af417d7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.922274] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327630, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.931351] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for the task: (returnval){ [ 837.931351] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52311803-550a-d3f4-21d4-b73e4d08fa7c" [ 837.931351] env[63293]: _type = "Task" [ 837.931351] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.944358] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52311803-550a-d3f4-21d4-b73e4d08fa7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.019201] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327628, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.022779] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d779c8e4-3f95-4f4a-b230-5b5278e8a8b7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.031648] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6bece6b-a5a0-40a8-bb2b-ea85dc653fe2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.065520] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ccafab-b5d7-4def-ad33-b3cc8aaf9d03 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.074241] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c1be09-65b1-4737-8efc-cb06dc57c2f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.090033] env[63293]: DEBUG nova.compute.provider_tree [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.107932] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327629, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.160297] env[63293]: INFO nova.compute.manager [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Took 24.25 seconds to build instance. [ 838.162591] env[63293]: DEBUG nova.network.neutron [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Updating instance_info_cache with network_info: [{"id": "20da136a-9449-4b46-be97-b9be3c4c2bfe", "address": "fa:16:3e:67:84:4b", "network": {"id": "d804be0f-03d3-4c34-afe5-a7fcafb986db", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-2128543466-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e5d75a3118c4a378f3a1ac529f5c5dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20da136a-94", "ovs_interfaceid": "20da136a-9449-4b46-be97-b9be3c4c2bfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.323218] env[63293]: INFO nova.compute.manager [-] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Took 1.75 seconds to deallocate network for instance. [ 838.418576] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327630, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.445568] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52311803-550a-d3f4-21d4-b73e4d08fa7c, 'name': SearchDatastore_Task, 'duration_secs': 0.021366} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.445722] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f08c1f72-fa32-4163-92b0-329b1003cb6e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.452165] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for the task: (returnval){ [ 838.452165] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f43b40-70ae-8917-c559-ec9adac5ed26" [ 838.452165] env[63293]: _type = "Task" [ 838.452165] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.461726] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f43b40-70ae-8917-c559-ec9adac5ed26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.514482] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327628, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523062} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.514802] env[63293]: INFO nova.virt.vmwareapi.ds_util [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 48603812-c3cc-4dae-b8ba-99b9ac5f7969/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk. [ 838.515742] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e33b5c0-b122-4f4a-b38f-67c6792dcee0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.546749] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 48603812-c3cc-4dae-b8ba-99b9ac5f7969/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.547176] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1abe4a5-7b9b-4a9c-abef-83eba5af308f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.563536] env[63293]: DEBUG nova.compute.manager [req-94d2095b-2986-4a76-95b7-89e8acba9abe req-e0f3a76b-bbb3-4423-858c-f60ec62b5999 service nova] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Received event network-vif-deleted-0c416db6-ba22-41a5-b45a-cf9c3f6b1624 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.563779] env[63293]: DEBUG nova.compute.manager [req-94d2095b-2986-4a76-95b7-89e8acba9abe req-e0f3a76b-bbb3-4423-858c-f60ec62b5999 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Received event network-changed-0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.563950] env[63293]: DEBUG nova.compute.manager [req-94d2095b-2986-4a76-95b7-89e8acba9abe req-e0f3a76b-bbb3-4423-858c-f60ec62b5999 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Refreshing instance network info cache due to event network-changed-0576b708-1b9b-4159-8079-9b97d8b4e70d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.564188] env[63293]: DEBUG oslo_concurrency.lockutils [req-94d2095b-2986-4a76-95b7-89e8acba9abe req-e0f3a76b-bbb3-4423-858c-f60ec62b5999 service nova] Acquiring lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.564336] env[63293]: DEBUG oslo_concurrency.lockutils [req-94d2095b-2986-4a76-95b7-89e8acba9abe req-e0f3a76b-bbb3-4423-858c-f60ec62b5999 service nova] Acquired lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.564572] env[63293]: DEBUG nova.network.neutron [req-94d2095b-2986-4a76-95b7-89e8acba9abe req-e0f3a76b-bbb3-4423-858c-f60ec62b5999 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Refreshing network info cache for port 0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.571184] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 838.571184] env[63293]: value = "task-1327631" [ 838.571184] env[63293]: _type = "Task" [ 838.571184] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.582154] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327631, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.595552] env[63293]: DEBUG nova.scheduler.client.report [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.613383] env[63293]: DEBUG oslo_vmware.api [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327629, 'name': PowerOnVM_Task, 'duration_secs': 0.635315} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.613492] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.613663] env[63293]: DEBUG nova.compute.manager [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.614496] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d0c278-d42b-4662-98a6-c363a1937c1b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.633356] env[63293]: DEBUG nova.compute.manager [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.667395] env[63293]: DEBUG nova.virt.hardware [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.667702] env[63293]: DEBUG nova.virt.hardware [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.667867] env[63293]: DEBUG nova.virt.hardware [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.668207] env[63293]: DEBUG nova.virt.hardware [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.668391] env[63293]: DEBUG nova.virt.hardware [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.668548] env[63293]: DEBUG nova.virt.hardware [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.668770] env[63293]: DEBUG nova.virt.hardware [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.668938] env[63293]: DEBUG nova.virt.hardware [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.669205] env[63293]: DEBUG nova.virt.hardware [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.669390] env[63293]: DEBUG nova.virt.hardware [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.669571] env[63293]: DEBUG nova.virt.hardware [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.670093] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Releasing lock "refresh_cache-93080446-fd45-4281-bf6e-35d1d964f41e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.670543] env[63293]: DEBUG nova.compute.manager [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Instance network_info: |[{"id": "20da136a-9449-4b46-be97-b9be3c4c2bfe", "address": "fa:16:3e:67:84:4b", "network": {"id": "d804be0f-03d3-4c34-afe5-a7fcafb986db", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-2128543466-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e5d75a3118c4a378f3a1ac529f5c5dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20da136a-94", "ovs_interfaceid": "20da136a-9449-4b46-be97-b9be3c4c2bfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 838.670921] env[63293]: DEBUG oslo_concurrency.lockutils [None req-73785574-067e-4fb9-9a15-ebe85a0870bc tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lock "831d7d9b-364f-4c29-bda5-e6c38291a973" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.980s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.671733] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519a453b-375d-4ad4-99a8-19fceb458b6c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.675431] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:84:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7806fe18-2b89-4386-87b1-f22876f82af2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '20da136a-9449-4b46-be97-b9be3c4c2bfe', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 838.685661] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Creating folder: Project (6e5d75a3118c4a378f3a1ac529f5c5dc). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.688848] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0805d10-17d2-4577-98fc-57d2f747b0fc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.697885] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49771a51-799b-447b-8078-97acc6a57b1e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.704424] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Created folder: Project (6e5d75a3118c4a378f3a1ac529f5c5dc) in parent group-v283678. [ 838.704688] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Creating folder: Instances. Parent ref: group-v283743. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.705495] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0fef3d11-84ea-450f-ad50-19c28dfa65ff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.722632] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Created folder: Instances in parent group-v283743. [ 838.722905] env[63293]: DEBUG oslo.service.loopingcall [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.723164] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 838.723377] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a145689d-94bc-4534-a01f-dcbc84aa6edf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.749089] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 838.749089] env[63293]: value = "task-1327634" [ 838.749089] env[63293]: _type = "Task" [ 838.749089] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.756528] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327634, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.830810] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.842077] env[63293]: DEBUG nova.compute.manager [req-4fb3f458-7592-4d42-817d-3746622dec7d req-1a09f484-ddb9-46e4-a65a-743cf6687180 service nova] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Received event network-changed-20da136a-9449-4b46-be97-b9be3c4c2bfe {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.842077] env[63293]: DEBUG nova.compute.manager [req-4fb3f458-7592-4d42-817d-3746622dec7d req-1a09f484-ddb9-46e4-a65a-743cf6687180 service nova] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Refreshing instance network info cache due to event network-changed-20da136a-9449-4b46-be97-b9be3c4c2bfe. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.842077] env[63293]: DEBUG oslo_concurrency.lockutils [req-4fb3f458-7592-4d42-817d-3746622dec7d req-1a09f484-ddb9-46e4-a65a-743cf6687180 service nova] Acquiring lock "refresh_cache-93080446-fd45-4281-bf6e-35d1d964f41e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.842077] env[63293]: DEBUG oslo_concurrency.lockutils [req-4fb3f458-7592-4d42-817d-3746622dec7d req-1a09f484-ddb9-46e4-a65a-743cf6687180 service nova] Acquired lock "refresh_cache-93080446-fd45-4281-bf6e-35d1d964f41e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.842077] env[63293]: DEBUG nova.network.neutron [req-4fb3f458-7592-4d42-817d-3746622dec7d req-1a09f484-ddb9-46e4-a65a-743cf6687180 service nova] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Refreshing network info cache for port 20da136a-9449-4b46-be97-b9be3c4c2bfe {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.917529] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327630, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.818892} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.917807] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 1e01a75d-122d-4122-9fed-4164d64d4ee8/1e01a75d-122d-4122-9fed-4164d64d4ee8.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 838.918039] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.918316] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-954fa18a-09ee-4d33-b973-0fd77dd35cb4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.924785] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for the task: (returnval){ [ 838.924785] env[63293]: value = "task-1327635" [ 838.924785] env[63293]: _type = "Task" [ 838.924785] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.935810] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327635, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.962202] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f43b40-70ae-8917-c559-ec9adac5ed26, 'name': SearchDatastore_Task, 'duration_secs': 0.059543} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.962500] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.962915] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 4593cabd-1331-4ee1-99f5-46edeac56859/4593cabd-1331-4ee1-99f5-46edeac56859.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 838.963029] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ac97be2-b49a-4c91-a102-0a3b6c16240b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.969800] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for the task: (returnval){ [ 838.969800] env[63293]: value = "task-1327636" [ 838.969800] env[63293]: _type = "Task" [ 838.969800] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.977865] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327636, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.081734] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327631, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.101947] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.102643] env[63293]: DEBUG nova.compute.manager [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 839.110772] env[63293]: DEBUG oslo_concurrency.lockutils [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.267s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.111089] env[63293]: DEBUG nova.objects.instance [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lazy-loading 'resources' on Instance uuid 495cb1b0-9194-44e6-9f68-de62033ea17d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 839.135677] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.194281] env[63293]: DEBUG nova.compute.manager [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 839.260344] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327634, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.438318] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327635, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.235615} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.441194] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.442235] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be6643ca-dae5-406f-84a3-58e74627c140 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.476297] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 1e01a75d-122d-4122-9fed-4164d64d4ee8/1e01a75d-122d-4122-9fed-4164d64d4ee8.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.477091] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3209d62a-22a4-4f2a-a21a-91a7ad4774f3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.506405] env[63293]: DEBUG nova.network.neutron [req-94d2095b-2986-4a76-95b7-89e8acba9abe req-e0f3a76b-bbb3-4423-858c-f60ec62b5999 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updated VIF entry in instance network info cache for port 0576b708-1b9b-4159-8079-9b97d8b4e70d. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 839.507126] env[63293]: DEBUG nova.network.neutron [req-94d2095b-2986-4a76-95b7-89e8acba9abe req-e0f3a76b-bbb3-4423-858c-f60ec62b5999 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updating instance_info_cache with network_info: [{"id": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "address": "fa:16:3e:c0:e8:15", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0576b708-1b", "ovs_interfaceid": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.524024] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327636, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.524024] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for the task: (returnval){ [ 839.524024] env[63293]: value = "task-1327637" [ 839.524024] env[63293]: _type = "Task" [ 839.524024] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.584675] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327631, 'name': ReconfigVM_Task, 'duration_secs': 0.897232} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.585176] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 48603812-c3cc-4dae-b8ba-99b9ac5f7969/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.586271] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad9cf6b-d9a0-4eff-8be2-d3ec90f4469b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.616097] env[63293]: DEBUG nova.compute.utils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.624508] env[63293]: DEBUG nova.compute.manager [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 839.624508] env[63293]: DEBUG nova.network.neutron [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 839.628905] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e4b9b34-1ab7-4761-8963-5a24513db819 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.646672] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 839.646672] env[63293]: value = "task-1327638" [ 839.646672] env[63293]: _type = "Task" [ 839.646672] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.658416] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327638, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.717697] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.719540] env[63293]: DEBUG nova.policy [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ebf82e99427d4171a4d510f7d3c966d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1687c33183e74b6ba70e021b6879cb93', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.723275] env[63293]: DEBUG nova.network.neutron [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Successfully updated port: 8b8079ba-2afe-42c9-9536-3b8951b51e00 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 839.761866] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327634, 'name': CreateVM_Task, 'duration_secs': 0.795081} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.765228] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 839.766291] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.766589] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.767035] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 839.767603] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32d2fcca-36af-4911-83a2-9d9429dd28da {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.772933] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for the task: (returnval){ [ 839.772933] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5249c9f5-7f65-ab5a-dea5-d8d7a5a424c3" [ 839.772933] env[63293]: _type = "Task" [ 839.772933] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.784302] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5249c9f5-7f65-ab5a-dea5-d8d7a5a424c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.804192] env[63293]: DEBUG nova.network.neutron [req-4fb3f458-7592-4d42-817d-3746622dec7d req-1a09f484-ddb9-46e4-a65a-743cf6687180 service nova] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Updated VIF entry in instance network info cache for port 20da136a-9449-4b46-be97-b9be3c4c2bfe. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 839.804574] env[63293]: DEBUG nova.network.neutron [req-4fb3f458-7592-4d42-817d-3746622dec7d req-1a09f484-ddb9-46e4-a65a-743cf6687180 service nova] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Updating instance_info_cache with network_info: [{"id": "20da136a-9449-4b46-be97-b9be3c4c2bfe", "address": "fa:16:3e:67:84:4b", "network": {"id": "d804be0f-03d3-4c34-afe5-a7fcafb986db", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-2128543466-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e5d75a3118c4a378f3a1ac529f5c5dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20da136a-94", "ovs_interfaceid": "20da136a-9449-4b46-be97-b9be3c4c2bfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.991463] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327636, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.784452} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.991463] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 4593cabd-1331-4ee1-99f5-46edeac56859/4593cabd-1331-4ee1-99f5-46edeac56859.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 839.992297] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.992297] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dffa5627-3418-496f-9f60-371788731ca0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.007534] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for the task: (returnval){ [ 840.007534] env[63293]: value = "task-1327639" [ 840.007534] env[63293]: _type = "Task" [ 840.007534] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.016845] env[63293]: DEBUG oslo_concurrency.lockutils [req-94d2095b-2986-4a76-95b7-89e8acba9abe req-e0f3a76b-bbb3-4423-858c-f60ec62b5999 service nova] Releasing lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.017534] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327639, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.033423] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327637, 'name': ReconfigVM_Task, 'duration_secs': 0.421093} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.036319] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 1e01a75d-122d-4122-9fed-4164d64d4ee8/1e01a75d-122d-4122-9fed-4164d64d4ee8.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.037481] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d1cb14f0-7b18-4a07-a30d-2da2a7b8996c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.045196] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for the task: (returnval){ [ 840.045196] env[63293]: value = "task-1327640" [ 840.045196] env[63293]: _type = "Task" [ 840.045196] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.050128] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c7ba36-4ac7-4c29-b596-15787189be70 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.059484] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327640, 'name': Rename_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.062468] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be322700-527c-4493-87b2-05602b079b83 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.097902] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b41cb2-f894-41e4-9dcc-6f746dd254f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.105724] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a529c24-295a-4815-823e-32f1540cd657 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.119717] env[63293]: DEBUG nova.compute.provider_tree [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.125397] env[63293]: DEBUG nova.compute.manager [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 840.157834] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327638, 'name': ReconfigVM_Task, 'duration_secs': 0.290332} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.157834] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.157834] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aed8b16e-f652-4227-ad2f-97fdddc11f4a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.164275] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 840.164275] env[63293]: value = "task-1327641" [ 840.164275] env[63293]: _type = "Task" [ 840.164275] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.168081] env[63293]: DEBUG nova.network.neutron [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Successfully created port: 4dbccf6a-053e-4933-91bc-1b7af5a260be {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 840.177426] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327641, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.227333] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "refresh_cache-c26d6a3a-3643-4b52-870a-4e2ed901d908" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.227437] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "refresh_cache-c26d6a3a-3643-4b52-870a-4e2ed901d908" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.227587] env[63293]: DEBUG nova.network.neutron [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 840.284168] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5249c9f5-7f65-ab5a-dea5-d8d7a5a424c3, 'name': SearchDatastore_Task, 'duration_secs': 0.01214} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.284469] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.284696] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 840.284931] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.285096] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.285294] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 840.285563] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bcb43468-ccf6-4fdb-a885-25c2b42b1395 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.298223] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 840.298455] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 840.299284] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8155b88e-d958-47a2-80b8-32f5281d4f72 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.304575] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for the task: (returnval){ [ 840.304575] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528b3018-14fa-cfe0-0a75-c5e4492e27f5" [ 840.304575] env[63293]: _type = "Task" [ 840.304575] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.308299] env[63293]: DEBUG oslo_concurrency.lockutils [req-4fb3f458-7592-4d42-817d-3746622dec7d req-1a09f484-ddb9-46e4-a65a-743cf6687180 service nova] Releasing lock "refresh_cache-93080446-fd45-4281-bf6e-35d1d964f41e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.313339] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528b3018-14fa-cfe0-0a75-c5e4492e27f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.500951] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "33d71260-26f1-482c-b93b-5f8e27c545f3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.501358] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "33d71260-26f1-482c-b93b-5f8e27c545f3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.501509] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "33d71260-26f1-482c-b93b-5f8e27c545f3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.501691] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "33d71260-26f1-482c-b93b-5f8e27c545f3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.501940] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "33d71260-26f1-482c-b93b-5f8e27c545f3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.504491] env[63293]: INFO nova.compute.manager [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Terminating instance [ 840.506446] env[63293]: DEBUG nova.compute.manager [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.506674] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.507584] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208689b7-5564-45dd-a2a0-3d234c255d60 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.521236] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327639, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06982} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.521236] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.521236] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.521822] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824f0fc4-c331-473f-b5ad-682c53fa566d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.525019] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-311cb011-f5b7-4179-bfc2-a341b78a8479 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.547460] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 4593cabd-1331-4ee1-99f5-46edeac56859/4593cabd-1331-4ee1-99f5-46edeac56859.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.549579] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4a6a1cb-438a-4e2a-8977-bfe86a5016d4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.565302] env[63293]: DEBUG oslo_vmware.api [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 840.565302] env[63293]: value = "task-1327642" [ 840.565302] env[63293]: _type = "Task" [ 840.565302] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.576960] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327640, 'name': Rename_Task, 'duration_secs': 0.138347} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.580746] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.581181] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for the task: (returnval){ [ 840.581181] env[63293]: value = "task-1327643" [ 840.581181] env[63293]: _type = "Task" [ 840.581181] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.581401] env[63293]: DEBUG oslo_vmware.api [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327642, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.581621] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a815512f-1644-413e-8ed4-0cedb16f8332 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.592516] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327643, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.593778] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for the task: (returnval){ [ 840.593778] env[63293]: value = "task-1327644" [ 840.593778] env[63293]: _type = "Task" [ 840.593778] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.602052] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327644, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.623530] env[63293]: DEBUG nova.scheduler.client.report [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.673558] env[63293]: DEBUG nova.compute.manager [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Received event network-changed-52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.673780] env[63293]: DEBUG nova.compute.manager [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Refreshing instance network info cache due to event network-changed-52a5b959-25ef-48ed-ad04-82f85a131929. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 840.674080] env[63293]: DEBUG oslo_concurrency.lockutils [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] Acquiring lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.674265] env[63293]: DEBUG oslo_concurrency.lockutils [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] Acquired lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.674440] env[63293]: DEBUG nova.network.neutron [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Refreshing network info cache for port 52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.686071] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327641, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.767399] env[63293]: DEBUG nova.network.neutron [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.816376] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528b3018-14fa-cfe0-0a75-c5e4492e27f5, 'name': SearchDatastore_Task, 'duration_secs': 0.02662} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.817605] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84fbdf7d-f835-4c8e-92e9-4025e2864585 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.822777] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for the task: (returnval){ [ 840.822777] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5244cc71-00dd-c22c-dd33-22cc19798553" [ 840.822777] env[63293]: _type = "Task" [ 840.822777] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.832549] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5244cc71-00dd-c22c-dd33-22cc19798553, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.985543] env[63293]: DEBUG nova.network.neutron [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Updating instance_info_cache with network_info: [{"id": "8b8079ba-2afe-42c9-9536-3b8951b51e00", "address": "fa:16:3e:cf:2c:55", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b8079ba-2a", "ovs_interfaceid": "8b8079ba-2afe-42c9-9536-3b8951b51e00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.080687] env[63293]: DEBUG oslo_vmware.api [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327642, 'name': PowerOffVM_Task, 'duration_secs': 0.386512} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.081024] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 841.081232] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 841.081540] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55f778f1-f491-41c4-9c00-824cda8f72cc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.092060] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327643, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.103538] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327644, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.128720] env[63293]: DEBUG oslo_concurrency.lockutils [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.018s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.130936] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.399s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.131190] env[63293]: DEBUG nova.objects.instance [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Lazy-loading 'resources' on Instance uuid c8009fc0-f7dc-4568-9909-7fa9f7f6a56c {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.136371] env[63293]: DEBUG nova.compute.manager [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 841.148683] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.148683] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.155510] env[63293]: INFO nova.scheduler.client.report [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Deleted allocations for instance 495cb1b0-9194-44e6-9f68-de62033ea17d [ 841.172668] env[63293]: DEBUG nova.virt.hardware [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.173016] env[63293]: DEBUG nova.virt.hardware [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.173229] env[63293]: DEBUG nova.virt.hardware [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.173534] env[63293]: DEBUG nova.virt.hardware [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.173783] env[63293]: DEBUG nova.virt.hardware [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.174100] env[63293]: DEBUG nova.virt.hardware [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.174354] env[63293]: DEBUG nova.virt.hardware [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.174525] env[63293]: DEBUG nova.virt.hardware [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.174727] env[63293]: DEBUG nova.virt.hardware [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.174938] env[63293]: DEBUG nova.virt.hardware [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.175172] env[63293]: DEBUG nova.virt.hardware [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.176592] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f46782a5-b766-4e59-910c-c749a043435e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.184764] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.184991] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.185192] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleting the datastore file [datastore1] 33d71260-26f1-482c-b93b-5f8e27c545f3 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.189522] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ef53eef-5744-405e-b01e-cc5f2db19fb2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.191454] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327641, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.195594] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de98b8f7-89af-4912-960e-9d497c2d90dc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.200896] env[63293]: DEBUG oslo_vmware.api [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 841.200896] env[63293]: value = "task-1327646" [ 841.200896] env[63293]: _type = "Task" [ 841.200896] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.218348] env[63293]: DEBUG oslo_vmware.api [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327646, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.333647] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5244cc71-00dd-c22c-dd33-22cc19798553, 'name': SearchDatastore_Task, 'duration_secs': 0.029484} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.333924] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.334239] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 93080446-fd45-4281-bf6e-35d1d964f41e/93080446-fd45-4281-bf6e-35d1d964f41e.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 841.334510] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a24bec01-d64f-4f32-96df-f1f6d0b808ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.341091] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for the task: (returnval){ [ 841.341091] env[63293]: value = "task-1327647" [ 841.341091] env[63293]: _type = "Task" [ 841.341091] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.349691] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327647, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.488926] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "refresh_cache-c26d6a3a-3643-4b52-870a-4e2ed901d908" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.489441] env[63293]: DEBUG nova.compute.manager [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Instance network_info: |[{"id": "8b8079ba-2afe-42c9-9536-3b8951b51e00", "address": "fa:16:3e:cf:2c:55", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b8079ba-2a", "ovs_interfaceid": "8b8079ba-2afe-42c9-9536-3b8951b51e00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 841.489910] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:2c:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ef746c57-cd18-4883-a0e9-c52937aaf41d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b8079ba-2afe-42c9-9536-3b8951b51e00', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.500043] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Creating folder: Project (7744835de0bc4e30b1b29ee5a439cd0d). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 841.500696] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a07fae2-8566-4aff-b09a-b68193c76501 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.512726] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Created folder: Project (7744835de0bc4e30b1b29ee5a439cd0d) in parent group-v283678. [ 841.513070] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Creating folder: Instances. Parent ref: group-v283746. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 841.513257] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8087f1d3-e06c-4927-b668-7dad7f5a5325 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.523146] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Created folder: Instances in parent group-v283746. [ 841.523462] env[63293]: DEBUG oslo.service.loopingcall [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.523787] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 841.524108] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff67c28c-69cd-45ac-8340-53144c442529 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.546982] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.546982] env[63293]: value = "task-1327650" [ 841.546982] env[63293]: _type = "Task" [ 841.546982] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.557536] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327650, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.595041] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327643, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.608739] env[63293]: DEBUG oslo_vmware.api [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327644, 'name': PowerOnVM_Task, 'duration_secs': 0.816623} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.608739] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.608739] env[63293]: INFO nova.compute.manager [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Took 10.43 seconds to spawn the instance on the hypervisor. [ 841.609148] env[63293]: DEBUG nova.compute.manager [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.609912] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dae53d9-0303-4c38-a8bd-d2725181e189 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.673841] env[63293]: DEBUG oslo_concurrency.lockutils [None req-233651bf-67e6-422d-bfce-21cc2947a688 tempest-SecurityGroupsTestJSON-1518094485 tempest-SecurityGroupsTestJSON-1518094485-project-member] Lock "495cb1b0-9194-44e6-9f68-de62033ea17d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.678s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.690179] env[63293]: DEBUG oslo_vmware.api [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327641, 'name': PowerOnVM_Task, 'duration_secs': 1.457705} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.690755] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.698088] env[63293]: DEBUG nova.compute.manager [None req-ee491557-e8eb-4149-86f2-1b2c2bdd104a tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.702545] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0ebb6e-4c77-4b68-8d6e-f2bed60ca0d3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.718335] env[63293]: DEBUG oslo_vmware.api [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327646, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.356675} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.721334] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.721802] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 841.722125] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.722423] env[63293]: INFO nova.compute.manager [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Took 1.22 seconds to destroy the instance on the hypervisor. [ 841.722856] env[63293]: DEBUG oslo.service.loopingcall [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.726411] env[63293]: DEBUG nova.compute.manager [-] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.726734] env[63293]: DEBUG nova.network.neutron [-] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 841.776343] env[63293]: DEBUG nova.network.neutron [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updated VIF entry in instance network info cache for port 52a5b959-25ef-48ed-ad04-82f85a131929. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 841.776343] env[63293]: DEBUG nova.network.neutron [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updating instance_info_cache with network_info: [{"id": "52a5b959-25ef-48ed-ad04-82f85a131929", "address": "fa:16:3e:56:4e:4a", "network": {"id": "b6024f78-d9d9-4fd8-a71e-9f996ce4a30f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-959509061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76803a169e7d424ab6aefe0d772b7827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52a5b959-25", "ovs_interfaceid": "52a5b959-25ef-48ed-ad04-82f85a131929", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.859534] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327647, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.900486] env[63293]: DEBUG nova.network.neutron [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Successfully updated port: 4dbccf6a-053e-4933-91bc-1b7af5a260be {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 841.922789] env[63293]: DEBUG nova.compute.manager [req-a0d6eba3-1fc2-4892-b4ca-400d26baa3fe req-3bab216f-ecd2-4fa0-b81f-dfcb1f9b1810 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Received event network-vif-plugged-4dbccf6a-053e-4933-91bc-1b7af5a260be {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.923103] env[63293]: DEBUG oslo_concurrency.lockutils [req-a0d6eba3-1fc2-4892-b4ca-400d26baa3fe req-3bab216f-ecd2-4fa0-b81f-dfcb1f9b1810 service nova] Acquiring lock "9ddf4f23-1279-4cbd-8212-10f344060445-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.923348] env[63293]: DEBUG oslo_concurrency.lockutils [req-a0d6eba3-1fc2-4892-b4ca-400d26baa3fe req-3bab216f-ecd2-4fa0-b81f-dfcb1f9b1810 service nova] Lock "9ddf4f23-1279-4cbd-8212-10f344060445-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.923568] env[63293]: DEBUG oslo_concurrency.lockutils [req-a0d6eba3-1fc2-4892-b4ca-400d26baa3fe req-3bab216f-ecd2-4fa0-b81f-dfcb1f9b1810 service nova] Lock "9ddf4f23-1279-4cbd-8212-10f344060445-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.924047] env[63293]: DEBUG nova.compute.manager [req-a0d6eba3-1fc2-4892-b4ca-400d26baa3fe req-3bab216f-ecd2-4fa0-b81f-dfcb1f9b1810 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] No waiting events found dispatching network-vif-plugged-4dbccf6a-053e-4933-91bc-1b7af5a260be {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.924263] env[63293]: WARNING nova.compute.manager [req-a0d6eba3-1fc2-4892-b4ca-400d26baa3fe req-3bab216f-ecd2-4fa0-b81f-dfcb1f9b1810 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Received unexpected event network-vif-plugged-4dbccf6a-053e-4933-91bc-1b7af5a260be for instance with vm_state building and task_state spawning. [ 842.042508] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba39bcbd-c414-4fa5-9c89-d5338d63a379 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.053836] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10dc6097-f713-4c82-b2e6-046cbcefa7d8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.061963] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327650, 'name': CreateVM_Task, 'duration_secs': 0.43285} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.093952] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 842.097495] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.097674] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.098013] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.101792] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96375c92-bca1-4490-afe6-4aaba2dfb689 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.104584] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e37b752e-1611-4b01-a83b-8c844cca9ecb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.116020] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327643, 'name': ReconfigVM_Task, 'duration_secs': 1.026591} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.116020] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 4593cabd-1331-4ee1-99f5-46edeac56859/4593cabd-1331-4ee1-99f5-46edeac56859.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.116020] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45a5e195-3eb7-4742-be8e-cea5c838d47d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.119335] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c402396-3730-4df3-ad3a-fe6c102c8c92 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.125161] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 842.125161] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b08ee9-35b2-e35f-9dcf-6d60f128c6d5" [ 842.125161] env[63293]: _type = "Task" [ 842.125161] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.143574] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for the task: (returnval){ [ 842.143574] env[63293]: value = "task-1327651" [ 842.143574] env[63293]: _type = "Task" [ 842.143574] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.144291] env[63293]: DEBUG nova.compute.provider_tree [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.145642] env[63293]: INFO nova.compute.manager [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Took 26.24 seconds to build instance. [ 842.155664] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b08ee9-35b2-e35f-9dcf-6d60f128c6d5, 'name': SearchDatastore_Task, 'duration_secs': 0.018189} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.157246] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.157627] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.157984] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.158277] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.158626] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.163022] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae12a04e-074d-4adc-9066-3a0482b472d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.165171] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327651, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.172591] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.172806] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 842.173582] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-318ac202-dc30-422b-acee-d33f29481271 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.178924] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 842.178924] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d2788d-b3ac-2c54-9dcb-cba7a3de1dde" [ 842.178924] env[63293]: _type = "Task" [ 842.178924] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.187468] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d2788d-b3ac-2c54-9dcb-cba7a3de1dde, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.282452] env[63293]: DEBUG oslo_concurrency.lockutils [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] Releasing lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.285880] env[63293]: DEBUG nova.compute.manager [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Received event network-vif-plugged-8b8079ba-2afe-42c9-9536-3b8951b51e00 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.285880] env[63293]: DEBUG oslo_concurrency.lockutils [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] Acquiring lock "c26d6a3a-3643-4b52-870a-4e2ed901d908-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.285880] env[63293]: DEBUG oslo_concurrency.lockutils [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] Lock "c26d6a3a-3643-4b52-870a-4e2ed901d908-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.285880] env[63293]: DEBUG oslo_concurrency.lockutils [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] Lock "c26d6a3a-3643-4b52-870a-4e2ed901d908-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.285880] env[63293]: DEBUG nova.compute.manager [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] No waiting events found dispatching network-vif-plugged-8b8079ba-2afe-42c9-9536-3b8951b51e00 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 842.286199] env[63293]: WARNING nova.compute.manager [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Received unexpected event network-vif-plugged-8b8079ba-2afe-42c9-9536-3b8951b51e00 for instance with vm_state building and task_state spawning. [ 842.286199] env[63293]: DEBUG nova.compute.manager [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Received event network-changed-8b8079ba-2afe-42c9-9536-3b8951b51e00 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.286199] env[63293]: DEBUG nova.compute.manager [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Refreshing instance network info cache due to event network-changed-8b8079ba-2afe-42c9-9536-3b8951b51e00. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 842.286199] env[63293]: DEBUG oslo_concurrency.lockutils [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] Acquiring lock "refresh_cache-c26d6a3a-3643-4b52-870a-4e2ed901d908" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.286199] env[63293]: DEBUG oslo_concurrency.lockutils [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] Acquired lock "refresh_cache-c26d6a3a-3643-4b52-870a-4e2ed901d908" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.286367] env[63293]: DEBUG nova.network.neutron [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Refreshing network info cache for port 8b8079ba-2afe-42c9-9536-3b8951b51e00 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 842.356226] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327647, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.689247} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.356226] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 93080446-fd45-4281-bf6e-35d1d964f41e/93080446-fd45-4281-bf6e-35d1d964f41e.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 842.356226] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 842.356226] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e6e96baa-6b3a-420c-9d17-2054f68ed320 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.362030] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for the task: (returnval){ [ 842.362030] env[63293]: value = "task-1327652" [ 842.362030] env[63293]: _type = "Task" [ 842.362030] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.375484] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327652, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.406294] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.406488] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.406542] env[63293]: DEBUG nova.network.neutron [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.606046] env[63293]: DEBUG nova.network.neutron [-] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.648138] env[63293]: DEBUG nova.scheduler.client.report [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.652575] env[63293]: DEBUG oslo_concurrency.lockutils [None req-80824312-0551-46fd-91da-bb4080e700b9 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Lock "1e01a75d-122d-4122-9fed-4164d64d4ee8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.984s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.664891] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327651, 'name': Rename_Task, 'duration_secs': 0.303521} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.665230] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.665412] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4e2305d-efbf-45ef-805e-406658ef5df1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.674034] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for the task: (returnval){ [ 842.674034] env[63293]: value = "task-1327653" [ 842.674034] env[63293]: _type = "Task" [ 842.674034] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.685996] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327653, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.691657] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d2788d-b3ac-2c54-9dcb-cba7a3de1dde, 'name': SearchDatastore_Task, 'duration_secs': 0.009014} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.692459] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c501f124-be85-4acb-b34f-dde68326825f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.697884] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 842.697884] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f0fb66-c191-75c5-def2-b7865c23ed02" [ 842.697884] env[63293]: _type = "Task" [ 842.697884] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.706226] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f0fb66-c191-75c5-def2-b7865c23ed02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.871897] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327652, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075842} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.872132] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 842.872986] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9679c8-b2fe-4ebf-8ade-0d67e9701fe9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.895300] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 93080446-fd45-4281-bf6e-35d1d964f41e/93080446-fd45-4281-bf6e-35d1d964f41e.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 842.896516] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbd41b6e-38bd-42bd-a4f2-bcff6f8e272e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.922020] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for the task: (returnval){ [ 842.922020] env[63293]: value = "task-1327654" [ 842.922020] env[63293]: _type = "Task" [ 842.922020] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.928539] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327654, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.965131] env[63293]: DEBUG nova.network.neutron [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.043874] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquiring lock "1e01a75d-122d-4122-9fed-4164d64d4ee8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.044170] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Lock "1e01a75d-122d-4122-9fed-4164d64d4ee8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.044473] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquiring lock "1e01a75d-122d-4122-9fed-4164d64d4ee8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.044573] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Lock "1e01a75d-122d-4122-9fed-4164d64d4ee8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.044751] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Lock "1e01a75d-122d-4122-9fed-4164d64d4ee8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.046788] env[63293]: INFO nova.compute.manager [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Terminating instance [ 843.048699] env[63293]: DEBUG nova.compute.manager [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 843.048905] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 843.050047] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ce9c84-c4da-4716-a968-8e8182877e6b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.062165] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 843.062165] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0c3699d-aef2-45f8-8c6b-ae8ae3052b24 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.070274] env[63293]: DEBUG oslo_vmware.api [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for the task: (returnval){ [ 843.070274] env[63293]: value = "task-1327655" [ 843.070274] env[63293]: _type = "Task" [ 843.070274] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.077184] env[63293]: DEBUG nova.compute.manager [req-8b33938c-6c87-4e5d-9e59-f49146ff232a req-628f071c-0929-4001-85af-ea9b591704b9 service nova] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Received event network-vif-deleted-01c58487-3d79-41c6-be70-8e4c5f2efe50 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.081134] env[63293]: DEBUG oslo_vmware.api [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327655, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.108908] env[63293]: INFO nova.compute.manager [-] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Took 1.38 seconds to deallocate network for instance. [ 843.136193] env[63293]: DEBUG nova.network.neutron [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Updated VIF entry in instance network info cache for port 8b8079ba-2afe-42c9-9536-3b8951b51e00. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 843.136625] env[63293]: DEBUG nova.network.neutron [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Updating instance_info_cache with network_info: [{"id": "8b8079ba-2afe-42c9-9536-3b8951b51e00", "address": "fa:16:3e:cf:2c:55", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b8079ba-2a", "ovs_interfaceid": "8b8079ba-2afe-42c9-9536-3b8951b51e00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.152589] env[63293]: DEBUG nova.network.neutron [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance_info_cache with network_info: [{"id": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "address": "fa:16:3e:93:d3:25", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dbccf6a-05", "ovs_interfaceid": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.159283] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.028s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.163031] env[63293]: DEBUG nova.compute.manager [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.165127] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.370s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.166523] env[63293]: INFO nova.compute.claims [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.189094] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327653, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.194905] env[63293]: INFO nova.scheduler.client.report [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Deleted allocations for instance c8009fc0-f7dc-4568-9909-7fa9f7f6a56c [ 843.211277] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f0fb66-c191-75c5-def2-b7865c23ed02, 'name': SearchDatastore_Task, 'duration_secs': 0.02806} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.211573] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.211832] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] c26d6a3a-3643-4b52-870a-4e2ed901d908/c26d6a3a-3643-4b52-870a-4e2ed901d908.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 843.212134] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fa2795b7-242e-42f0-a04c-174703ea71cd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.221395] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 843.221395] env[63293]: value = "task-1327656" [ 843.221395] env[63293]: _type = "Task" [ 843.221395] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.231157] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327656, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.430393] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327654, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.580035] env[63293]: DEBUG oslo_vmware.api [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327655, 'name': PowerOffVM_Task, 'duration_secs': 0.197272} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.580429] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.580629] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 843.580901] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-93c7c4ee-c324-48aa-a63e-49203a10ae42 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.615913] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.639675] env[63293]: DEBUG oslo_concurrency.lockutils [req-cfee9a97-0fe8-4435-ac81-ed48ab8d0e59 req-20fa88bb-5d64-4b98-b586-cc1d6fa06d22 service nova] Releasing lock "refresh_cache-c26d6a3a-3643-4b52-870a-4e2ed901d908" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.655650] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.656046] env[63293]: DEBUG nova.compute.manager [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Instance network_info: |[{"id": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "address": "fa:16:3e:93:d3:25", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dbccf6a-05", "ovs_interfaceid": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 843.656546] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:d3:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4dbccf6a-053e-4933-91bc-1b7af5a260be', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.664832] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating folder: Project (1687c33183e74b6ba70e021b6879cb93). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.666401] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-22a73727-de70-4839-a25e-d619bdab8ccc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.671199] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 843.671506] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 843.671621] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Deleting the datastore file [datastore1] 1e01a75d-122d-4122-9fed-4164d64d4ee8 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.680022] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-283af5fb-6e27-498a-b5fa-14487303c490 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.689953] env[63293]: DEBUG oslo_vmware.api [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for the task: (returnval){ [ 843.689953] env[63293]: value = "task-1327659" [ 843.689953] env[63293]: _type = "Task" [ 843.689953] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.694699] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Created folder: Project (1687c33183e74b6ba70e021b6879cb93) in parent group-v283678. [ 843.694920] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating folder: Instances. Parent ref: group-v283749. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.695689] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327653, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.699719] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2957933-aa50-46d0-9949-78be37f260ef {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.702214] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.708956] env[63293]: DEBUG oslo_vmware.api [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327659, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.710690] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55abb3e3-52cc-40fd-8602-4fb72477767a tempest-ServerDiagnosticsTest-883464061 tempest-ServerDiagnosticsTest-883464061-project-member] Lock "c8009fc0-f7dc-4568-9909-7fa9f7f6a56c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.454s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.711738] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Created folder: Instances in parent group-v283749. [ 843.711976] env[63293]: DEBUG oslo.service.loopingcall [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.712376] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 843.712607] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2f9e38aa-3fae-4785-8699-3f5db5de630a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.738934] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327656, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.740405] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.740405] env[63293]: value = "task-1327661" [ 843.740405] env[63293]: _type = "Task" [ 843.740405] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.749091] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327661, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.934383] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327654, 'name': ReconfigVM_Task, 'duration_secs': 0.711544} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.935022] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 93080446-fd45-4281-bf6e-35d1d964f41e/93080446-fd45-4281-bf6e-35d1d964f41e.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 843.935900] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-912f3c0f-d940-4be5-aee5-7e0ff621ed1f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.945426] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for the task: (returnval){ [ 843.945426] env[63293]: value = "task-1327662" [ 843.945426] env[63293]: _type = "Task" [ 843.945426] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.962296] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327662, 'name': Rename_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.177867] env[63293]: DEBUG nova.compute.manager [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Received event network-changed-4dbccf6a-053e-4933-91bc-1b7af5a260be {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.178639] env[63293]: DEBUG nova.compute.manager [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Refreshing instance network info cache due to event network-changed-4dbccf6a-053e-4933-91bc-1b7af5a260be. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.179645] env[63293]: DEBUG oslo_concurrency.lockutils [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] Acquiring lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.180218] env[63293]: DEBUG oslo_concurrency.lockutils [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] Acquired lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.181405] env[63293]: DEBUG nova.network.neutron [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Refreshing network info cache for port 4dbccf6a-053e-4933-91bc-1b7af5a260be {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.210885] env[63293]: DEBUG oslo_vmware.api [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327653, 'name': PowerOnVM_Task, 'duration_secs': 1.505184} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.215118] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 844.215442] env[63293]: INFO nova.compute.manager [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Took 10.60 seconds to spawn the instance on the hypervisor. [ 844.216028] env[63293]: DEBUG nova.compute.manager [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.217483] env[63293]: DEBUG oslo_vmware.api [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Task: {'id': task-1327659, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.322832} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.218279] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c793a8c-b6d3-4f8c-ae0d-0ed344a397b0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.221436] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 844.221638] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 844.221821] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 844.221998] env[63293]: INFO nova.compute.manager [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Took 1.17 seconds to destroy the instance on the hypervisor. [ 844.222274] env[63293]: DEBUG oslo.service.loopingcall [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.222491] env[63293]: DEBUG nova.compute.manager [-] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.222616] env[63293]: DEBUG nova.network.neutron [-] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 844.253219] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327656, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.747699} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.256502] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] c26d6a3a-3643-4b52-870a-4e2ed901d908/c26d6a3a-3643-4b52-870a-4e2ed901d908.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 844.256734] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.256996] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327661, 'name': CreateVM_Task, 'duration_secs': 0.501554} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.257227] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bda70ec7-d5ff-44a0-883c-8526c03a7239 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.259092] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 844.262051] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.262229] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.262549] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.263644] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe88d137-79fc-49c9-8b1f-4bcf95b720bf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.269556] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 844.269556] env[63293]: value = "task-1327663" [ 844.269556] env[63293]: _type = "Task" [ 844.269556] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.274382] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 844.274382] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5206fbaa-0c13-48c2-e272-8449b860c6b3" [ 844.274382] env[63293]: _type = "Task" [ 844.274382] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.287766] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327663, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.293900] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5206fbaa-0c13-48c2-e272-8449b860c6b3, 'name': SearchDatastore_Task, 'duration_secs': 0.013669} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.296803] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.297061] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.297289] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.297445] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.297663] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.298141] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-444efe7d-7116-41e6-9b0e-b24d114c0209 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.306902] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.307133] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.307905] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c057580-66b2-43f9-bebb-00113af8b640 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.316261] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 844.316261] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520920a3-feff-bd7c-2dd4-14a2e7e8537d" [ 844.316261] env[63293]: _type = "Task" [ 844.316261] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.324697] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520920a3-feff-bd7c-2dd4-14a2e7e8537d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.455282] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327662, 'name': Rename_Task, 'duration_secs': 0.257812} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.459424] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 844.459424] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36f8bd7a-4486-47a4-8be9-476baa255645 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.464584] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for the task: (returnval){ [ 844.464584] env[63293]: value = "task-1327664" [ 844.464584] env[63293]: _type = "Task" [ 844.464584] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.473908] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.606108] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8902eae-bdcc-4fd5-a732-db3b910cdae5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.615044] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51db114-8676-4d09-a8e4-0bdf95b5f56d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.645801] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e60881-266c-4601-97d6-54946a576ed7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.654076] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf8e9a0-a398-45a9-86b2-15b489d4a2d3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.670038] env[63293]: DEBUG nova.compute.provider_tree [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.754694] env[63293]: INFO nova.compute.manager [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Took 27.72 seconds to build instance. [ 844.780131] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327663, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068423} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.780471] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 844.781270] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95602cc6-1e75-426f-915a-2381f43d394a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.803343] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] c26d6a3a-3643-4b52-870a-4e2ed901d908/c26d6a3a-3643-4b52-870a-4e2ed901d908.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 844.805747] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c38a11c-d7b5-442a-8235-85028e36c0d6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.830389] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520920a3-feff-bd7c-2dd4-14a2e7e8537d, 'name': SearchDatastore_Task, 'duration_secs': 0.01199} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.832987] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 844.832987] env[63293]: value = "task-1327665" [ 844.832987] env[63293]: _type = "Task" [ 844.832987] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.833204] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71004acd-ac69-4124-a248-cbea19ff6216 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.842810] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 844.842810] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5291afd0-c77d-b339-7d39-a21e4aef9c64" [ 844.842810] env[63293]: _type = "Task" [ 844.842810] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.847602] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327665, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.856321] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5291afd0-c77d-b339-7d39-a21e4aef9c64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.976257] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.013209] env[63293]: DEBUG nova.network.neutron [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updated VIF entry in instance network info cache for port 4dbccf6a-053e-4933-91bc-1b7af5a260be. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.013338] env[63293]: DEBUG nova.network.neutron [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance_info_cache with network_info: [{"id": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "address": "fa:16:3e:93:d3:25", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dbccf6a-05", "ovs_interfaceid": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.015495] env[63293]: DEBUG nova.network.neutron [-] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.176104] env[63293]: DEBUG nova.scheduler.client.report [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.201189] env[63293]: DEBUG nova.compute.manager [req-e0a09df3-1f59-4421-ba8c-9f1cba76d4d4 req-d28326c4-9ab4-4355-8855-97249b15b9b9 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Received event network-changed-26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.201189] env[63293]: DEBUG nova.compute.manager [req-e0a09df3-1f59-4421-ba8c-9f1cba76d4d4 req-d28326c4-9ab4-4355-8855-97249b15b9b9 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Refreshing instance network info cache due to event network-changed-26b86f34-f9b6-4b0b-abdd-d8acd51085fd. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 845.201189] env[63293]: DEBUG oslo_concurrency.lockutils [req-e0a09df3-1f59-4421-ba8c-9f1cba76d4d4 req-d28326c4-9ab4-4355-8855-97249b15b9b9 service nova] Acquiring lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.201189] env[63293]: DEBUG oslo_concurrency.lockutils [req-e0a09df3-1f59-4421-ba8c-9f1cba76d4d4 req-d28326c4-9ab4-4355-8855-97249b15b9b9 service nova] Acquired lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.201189] env[63293]: DEBUG nova.network.neutron [req-e0a09df3-1f59-4421-ba8c-9f1cba76d4d4 req-d28326c4-9ab4-4355-8855-97249b15b9b9 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Refreshing network info cache for port 26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 845.256495] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0923cee-3263-4be2-8b06-7ad7b989feb0 tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Lock "4593cabd-1331-4ee1-99f5-46edeac56859" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.191s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.344777] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327665, 'name': ReconfigVM_Task, 'duration_secs': 0.269333} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.345184] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Reconfigured VM instance instance-00000044 to attach disk [datastore2] c26d6a3a-3643-4b52-870a-4e2ed901d908/c26d6a3a-3643-4b52-870a-4e2ed901d908.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.345856] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-97434485-9e56-4015-bffe-42fe5ef46ead {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.358860] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5291afd0-c77d-b339-7d39-a21e4aef9c64, 'name': SearchDatastore_Task, 'duration_secs': 0.020266} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.360559] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.360844] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 9ddf4f23-1279-4cbd-8212-10f344060445/9ddf4f23-1279-4cbd-8212-10f344060445.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.361211] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 845.361211] env[63293]: value = "task-1327666" [ 845.361211] env[63293]: _type = "Task" [ 845.361211] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.361417] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8029f5d2-7dee-4dbd-b8a6-60f692908f58 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.372330] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327666, 'name': Rename_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.373737] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 845.373737] env[63293]: value = "task-1327667" [ 845.373737] env[63293]: _type = "Task" [ 845.373737] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.385875] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327667, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.484235] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.516132] env[63293]: DEBUG oslo_concurrency.lockutils [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] Releasing lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.516440] env[63293]: DEBUG nova.compute.manager [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Received event network-changed-26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.516619] env[63293]: DEBUG nova.compute.manager [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Refreshing instance network info cache due to event network-changed-26b86f34-f9b6-4b0b-abdd-d8acd51085fd. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 845.516816] env[63293]: DEBUG oslo_concurrency.lockutils [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] Acquiring lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.517291] env[63293]: INFO nova.compute.manager [-] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Took 1.29 seconds to deallocate network for instance. [ 845.683659] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.684222] env[63293]: DEBUG nova.compute.manager [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.687174] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.506s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.689309] env[63293]: INFO nova.compute.claims [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.760370] env[63293]: DEBUG nova.compute.manager [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 845.876415] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327666, 'name': Rename_Task, 'duration_secs': 0.160286} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.881682] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 845.882038] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5108d69-788b-4922-af74-56c2f8d08a2d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.890294] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327667, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.892037] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 845.892037] env[63293]: value = "task-1327668" [ 845.892037] env[63293]: _type = "Task" [ 845.892037] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.902035] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327668, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.979053] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.025343] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.028720] env[63293]: DEBUG nova.network.neutron [req-e0a09df3-1f59-4421-ba8c-9f1cba76d4d4 req-d28326c4-9ab4-4355-8855-97249b15b9b9 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updated VIF entry in instance network info cache for port 26b86f34-f9b6-4b0b-abdd-d8acd51085fd. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 846.029011] env[63293]: DEBUG nova.network.neutron [req-e0a09df3-1f59-4421-ba8c-9f1cba76d4d4 req-d28326c4-9ab4-4355-8855-97249b15b9b9 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updating instance_info_cache with network_info: [{"id": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "address": "fa:16:3e:fc:c1:7d", "network": {"id": "bbf31454-8e46-45d9-b10a-fccda17654c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1875587777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "6fa6cd8dea16412db105faf2ff93a6ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26b86f34-f9", "ovs_interfaceid": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.194378] env[63293]: DEBUG nova.compute.utils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.199668] env[63293]: DEBUG nova.compute.manager [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.199668] env[63293]: DEBUG nova.network.neutron [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.243075] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.243383] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.244918] env[63293]: DEBUG nova.policy [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3abe71ab8b024338924c7f6389441a93', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d72964783a4d0e95daa2305c6148cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.296911] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.388321] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327667, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.676383} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.388742] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 9ddf4f23-1279-4cbd-8212-10f344060445/9ddf4f23-1279-4cbd-8212-10f344060445.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.389024] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.389368] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-995447db-0e0d-4e9e-8190-957b7341defc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.399935] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 846.399935] env[63293]: value = "task-1327669" [ 846.399935] env[63293]: _type = "Task" [ 846.399935] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.407188] env[63293]: DEBUG oslo_vmware.api [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327668, 'name': PowerOnVM_Task, 'duration_secs': 0.489933} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.408265] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 846.408265] env[63293]: INFO nova.compute.manager [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Took 7.77 seconds to spawn the instance on the hypervisor. [ 846.408476] env[63293]: DEBUG nova.compute.manager [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.409365] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213bb127-84e1-4f9b-a8a3-a7a93d8cc4b9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.415769] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327669, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.478097] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.536736] env[63293]: DEBUG oslo_concurrency.lockutils [req-e0a09df3-1f59-4421-ba8c-9f1cba76d4d4 req-d28326c4-9ab4-4355-8855-97249b15b9b9 service nova] Releasing lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.537247] env[63293]: DEBUG oslo_concurrency.lockutils [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] Acquired lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.537493] env[63293]: DEBUG nova.network.neutron [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Refreshing network info cache for port 26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.699490] env[63293]: DEBUG nova.compute.manager [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.770357] env[63293]: DEBUG nova.compute.manager [req-8aadec52-45eb-4cf7-8a3a-742576ebf52e req-5fe177f6-6d55-48b3-9a9f-460b3e1fb07d service nova] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Received event network-vif-deleted-1582d90f-11d5-4ec4-8173-e1fcb11d3bc8 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.863676] env[63293]: DEBUG nova.network.neutron [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Successfully created port: b6279a11-248c-4e66-9df7-aa175821b8f1 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.916582] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327669, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071067} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.916582] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.917256] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac61e42-11d9-4fd8-a009-2e4f55c11128 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.949044] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 9ddf4f23-1279-4cbd-8212-10f344060445/9ddf4f23-1279-4cbd-8212-10f344060445.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.954057] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-862cd04f-9425-44d2-a37e-18e62458c9af {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.971155] env[63293]: INFO nova.compute.manager [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Took 25.43 seconds to build instance. [ 846.982882] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.984842] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 846.984842] env[63293]: value = "task-1327670" [ 846.984842] env[63293]: _type = "Task" [ 846.984842] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.991989] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327670, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.204561] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8ec103-c920-4cdc-af5c-f274799fef76 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.215741] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddfb8237-dd33-4625-850b-832ad46e72bb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.254227] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eccaa818-f63d-4519-b0b0-d74a7cf3ce31 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.263494] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c53c11a-0548-4f84-8102-5b01f678f7b9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.279673] env[63293]: DEBUG nova.compute.provider_tree [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.312066] env[63293]: DEBUG nova.network.neutron [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Successfully created port: 31fd0cd3-702f-4cea-860d-fc21f02effb2 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.391314] env[63293]: DEBUG nova.network.neutron [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updated VIF entry in instance network info cache for port 26b86f34-f9b6-4b0b-abdd-d8acd51085fd. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 847.391314] env[63293]: DEBUG nova.network.neutron [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updating instance_info_cache with network_info: [{"id": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "address": "fa:16:3e:fc:c1:7d", "network": {"id": "bbf31454-8e46-45d9-b10a-fccda17654c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1875587777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "6fa6cd8dea16412db105faf2ff93a6ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26b86f34-f9", "ovs_interfaceid": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.396586] env[63293]: DEBUG nova.compute.manager [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Received event network-changed-26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.396586] env[63293]: DEBUG nova.compute.manager [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Refreshing instance network info cache due to event network-changed-26b86f34-f9b6-4b0b-abdd-d8acd51085fd. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.396586] env[63293]: DEBUG oslo_concurrency.lockutils [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] Acquiring lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.476522] env[63293]: DEBUG oslo_concurrency.lockutils [None req-59a703b8-9c4b-4ada-9609-2d202f87379a tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "c26d6a3a-3643-4b52-870a-4e2ed901d908" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.342s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.481884] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.494225] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327670, 'name': ReconfigVM_Task, 'duration_secs': 0.283606} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.494835] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 9ddf4f23-1279-4cbd-8212-10f344060445/9ddf4f23-1279-4cbd-8212-10f344060445.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.495110] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2f20d76c-8c77-46cd-a97e-4b56c043d09b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.501181] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 847.501181] env[63293]: value = "task-1327671" [ 847.501181] env[63293]: _type = "Task" [ 847.501181] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.510316] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327671, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.722422] env[63293]: DEBUG nova.compute.manager [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.747984] env[63293]: DEBUG nova.virt.hardware [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.749207] env[63293]: DEBUG nova.virt.hardware [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.749207] env[63293]: DEBUG nova.virt.hardware [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.749207] env[63293]: DEBUG nova.virt.hardware [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.749207] env[63293]: DEBUG nova.virt.hardware [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.749207] env[63293]: DEBUG nova.virt.hardware [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.749458] env[63293]: DEBUG nova.virt.hardware [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.749820] env[63293]: DEBUG nova.virt.hardware [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.749927] env[63293]: DEBUG nova.virt.hardware [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.750045] env[63293]: DEBUG nova.virt.hardware [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.750284] env[63293]: DEBUG nova.virt.hardware [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.751247] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e199ef-6ed6-4daf-bd5b-6ef01e205d98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.761669] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb392d5-aa18-4472-b8d3-aaaf36f17740 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.784432] env[63293]: DEBUG nova.scheduler.client.report [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.897325] env[63293]: DEBUG oslo_concurrency.lockutils [req-29be3b66-7ed2-4573-a72e-b92f5f13437f req-dcd3e0b9-f0b2-41ff-80b5-76aec1f71759 service nova] Releasing lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.898252] env[63293]: DEBUG oslo_concurrency.lockutils [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] Acquired lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.898946] env[63293]: DEBUG nova.network.neutron [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Refreshing network info cache for port 26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.982879] env[63293]: DEBUG nova.compute.manager [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 847.985888] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.014439] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327671, 'name': Rename_Task, 'duration_secs': 0.134006} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.016920] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.017262] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef6abd9f-1808-4cc9-9c79-681dd5186d7c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.029775] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 848.029775] env[63293]: value = "task-1327672" [ 848.029775] env[63293]: _type = "Task" [ 848.029775] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.041934] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327672, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.290921] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.603s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.291064] env[63293]: DEBUG nova.compute.manager [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.294106] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.071s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.295533] env[63293]: INFO nova.compute.claims [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.484718] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.518117] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.544289] env[63293]: DEBUG oslo_vmware.api [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327672, 'name': PowerOnVM_Task, 'duration_secs': 0.463252} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.544575] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.544779] env[63293]: INFO nova.compute.manager [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Took 7.41 seconds to spawn the instance on the hypervisor. [ 848.544957] env[63293]: DEBUG nova.compute.manager [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.545765] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4ac9b6-f4c6-4f80-ae02-ffb55c1c0ec7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.634302] env[63293]: DEBUG nova.network.neutron [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updated VIF entry in instance network info cache for port 26b86f34-f9b6-4b0b-abdd-d8acd51085fd. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 848.634700] env[63293]: DEBUG nova.network.neutron [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updating instance_info_cache with network_info: [{"id": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "address": "fa:16:3e:fc:c1:7d", "network": {"id": "bbf31454-8e46-45d9-b10a-fccda17654c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1875587777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "6fa6cd8dea16412db105faf2ff93a6ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26b86f34-f9", "ovs_interfaceid": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.650440] env[63293]: DEBUG oslo_concurrency.lockutils [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "c26d6a3a-3643-4b52-870a-4e2ed901d908" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.650440] env[63293]: DEBUG oslo_concurrency.lockutils [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "c26d6a3a-3643-4b52-870a-4e2ed901d908" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.650655] env[63293]: DEBUG oslo_concurrency.lockutils [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "c26d6a3a-3643-4b52-870a-4e2ed901d908-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.650853] env[63293]: DEBUG oslo_concurrency.lockutils [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "c26d6a3a-3643-4b52-870a-4e2ed901d908-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.651021] env[63293]: DEBUG oslo_concurrency.lockutils [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "c26d6a3a-3643-4b52-870a-4e2ed901d908-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.653567] env[63293]: INFO nova.compute.manager [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Terminating instance [ 848.655384] env[63293]: DEBUG nova.compute.manager [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 848.655598] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 848.656687] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33421f6-aa69-4089-b588-d3496487dfb3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.665713] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 848.665968] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d25863f-47f3-4259-8bfc-7149847e3655 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.672530] env[63293]: DEBUG oslo_vmware.api [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 848.672530] env[63293]: value = "task-1327673" [ 848.672530] env[63293]: _type = "Task" [ 848.672530] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.681119] env[63293]: DEBUG oslo_vmware.api [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327673, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.736924] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquiring lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.737273] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.737509] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquiring lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.737695] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.738311] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.740248] env[63293]: INFO nova.compute.manager [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Terminating instance [ 848.742278] env[63293]: DEBUG nova.compute.manager [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 848.742485] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 848.743328] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d04c10e-4022-47f2-8f6a-16f0cdcd045b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.751218] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 848.751482] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd8dc3bc-11af-4ac9-b321-a1720bb84fa0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.757597] env[63293]: DEBUG oslo_vmware.api [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 848.757597] env[63293]: value = "task-1327674" [ 848.757597] env[63293]: _type = "Task" [ 848.757597] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.765593] env[63293]: DEBUG oslo_vmware.api [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327674, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.801023] env[63293]: DEBUG nova.compute.utils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.806025] env[63293]: DEBUG nova.compute.manager [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 848.806025] env[63293]: DEBUG nova.network.neutron [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.856838] env[63293]: DEBUG nova.policy [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd2a9495ab0b4768a4075a71f25aa9a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff8479fc251544e9b4618deccf992754', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.959412] env[63293]: DEBUG nova.compute.manager [req-08381221-e46c-46fd-9ad2-5f3bf407588a req-45ee50a5-1f89-40a9-ab53-55a7671f9d97 service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Received event network-vif-plugged-b6279a11-248c-4e66-9df7-aa175821b8f1 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.959663] env[63293]: DEBUG oslo_concurrency.lockutils [req-08381221-e46c-46fd-9ad2-5f3bf407588a req-45ee50a5-1f89-40a9-ab53-55a7671f9d97 service nova] Acquiring lock "f427933b-dd2c-49a9-b401-a647635a1eab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.959885] env[63293]: DEBUG oslo_concurrency.lockutils [req-08381221-e46c-46fd-9ad2-5f3bf407588a req-45ee50a5-1f89-40a9-ab53-55a7671f9d97 service nova] Lock "f427933b-dd2c-49a9-b401-a647635a1eab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.960100] env[63293]: DEBUG oslo_concurrency.lockutils [req-08381221-e46c-46fd-9ad2-5f3bf407588a req-45ee50a5-1f89-40a9-ab53-55a7671f9d97 service nova] Lock "f427933b-dd2c-49a9-b401-a647635a1eab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.960305] env[63293]: DEBUG nova.compute.manager [req-08381221-e46c-46fd-9ad2-5f3bf407588a req-45ee50a5-1f89-40a9-ab53-55a7671f9d97 service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] No waiting events found dispatching network-vif-plugged-b6279a11-248c-4e66-9df7-aa175821b8f1 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 848.960480] env[63293]: WARNING nova.compute.manager [req-08381221-e46c-46fd-9ad2-5f3bf407588a req-45ee50a5-1f89-40a9-ab53-55a7671f9d97 service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Received unexpected event network-vif-plugged-b6279a11-248c-4e66-9df7-aa175821b8f1 for instance with vm_state building and task_state spawning. [ 848.984604] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327664, 'name': PowerOnVM_Task} progress is 81%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.067384] env[63293]: INFO nova.compute.manager [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Took 24.80 seconds to build instance. [ 849.095237] env[63293]: DEBUG nova.network.neutron [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Successfully updated port: b6279a11-248c-4e66-9df7-aa175821b8f1 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.105387] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquiring lock "4593cabd-1331-4ee1-99f5-46edeac56859" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.105656] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Lock "4593cabd-1331-4ee1-99f5-46edeac56859" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.105883] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquiring lock "4593cabd-1331-4ee1-99f5-46edeac56859-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.107452] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Lock "4593cabd-1331-4ee1-99f5-46edeac56859-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.107452] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Lock "4593cabd-1331-4ee1-99f5-46edeac56859-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.113753] env[63293]: INFO nova.compute.manager [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Terminating instance [ 849.115867] env[63293]: DEBUG nova.compute.manager [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 849.116539] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 849.117516] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ccdedd8-40c3-4c83-9090-eb3e1d12d353 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.123368] env[63293]: DEBUG nova.network.neutron [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Successfully created port: ee3d149b-d66e-439b-a49e-6f2e6c9d9495 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.127905] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 849.128215] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25da5f56-376a-421f-a7d2-e606f9686523 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.135281] env[63293]: DEBUG oslo_vmware.api [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for the task: (returnval){ [ 849.135281] env[63293]: value = "task-1327675" [ 849.135281] env[63293]: _type = "Task" [ 849.135281] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.141222] env[63293]: DEBUG oslo_concurrency.lockutils [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] Releasing lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.141222] env[63293]: DEBUG nova.compute.manager [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Received event network-changed-26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.141509] env[63293]: DEBUG nova.compute.manager [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Refreshing instance network info cache due to event network-changed-26b86f34-f9b6-4b0b-abdd-d8acd51085fd. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.141509] env[63293]: DEBUG oslo_concurrency.lockutils [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] Acquiring lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.141613] env[63293]: DEBUG oslo_concurrency.lockutils [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] Acquired lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.141847] env[63293]: DEBUG nova.network.neutron [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Refreshing network info cache for port 26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.148208] env[63293]: DEBUG oslo_vmware.api [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327675, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.182844] env[63293]: DEBUG oslo_vmware.api [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327673, 'name': PowerOffVM_Task, 'duration_secs': 0.205381} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.183181] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 849.183373] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 849.183653] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af4c8e25-c02a-4cd4-bc65-cac45f2f028e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.253969] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 849.254220] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 849.254407] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleting the datastore file [datastore2] c26d6a3a-3643-4b52-870a-4e2ed901d908 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.254692] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed128a07-5dc5-4571-bec8-9d0e8d0baf2d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.265031] env[63293]: DEBUG oslo_vmware.api [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 849.265031] env[63293]: value = "task-1327677" [ 849.265031] env[63293]: _type = "Task" [ 849.265031] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.271887] env[63293]: DEBUG oslo_vmware.api [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327674, 'name': PowerOffVM_Task, 'duration_secs': 0.233966} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.272539] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 849.272719] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 849.273086] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-058b5abd-dfb0-40fa-8594-55ecffdd8d39 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.277256] env[63293]: DEBUG oslo_vmware.api [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327677, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.306276] env[63293]: DEBUG nova.compute.manager [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.483297] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 849.483560] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 849.483756] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Deleting the datastore file [datastore2] 48603812-c3cc-4dae-b8ba-99b9ac5f7969 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.484054] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0210364a-e572-49f9-8510-212414915986 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.490071] env[63293]: DEBUG oslo_vmware.api [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327664, 'name': PowerOnVM_Task, 'duration_secs': 4.905042} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.490805] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.491038] env[63293]: INFO nova.compute.manager [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Took 13.37 seconds to spawn the instance on the hypervisor. [ 849.491510] env[63293]: DEBUG nova.compute.manager [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.492438] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e94dd8-2270-47df-a81e-195be6a6d1ab {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.500240] env[63293]: DEBUG oslo_vmware.api [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for the task: (returnval){ [ 849.500240] env[63293]: value = "task-1327679" [ 849.500240] env[63293]: _type = "Task" [ 849.500240] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.514200] env[63293]: DEBUG oslo_vmware.api [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327679, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.570055] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7cf3dba8-a4b5-43b0-9c7a-f5c17b3f7da3 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.293s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.645912] env[63293]: DEBUG oslo_vmware.api [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327675, 'name': PowerOffVM_Task, 'duration_secs': 0.303293} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.648450] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 849.648632] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 849.650950] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba30c4ac-05d4-40d7-bb1f-5661f3af4025 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.694335] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef957d2-0d65-43a0-94f3-b00233530e39 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.705179] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-510cf0ab-8465-45f9-ac4b-56e469d769ea {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.742055] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919e4af3-6714-4eef-a4eb-eec236313df6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.745244] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 849.746548] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 849.746548] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Deleting the datastore file [datastore1] 4593cabd-1331-4ee1-99f5-46edeac56859 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.748582] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a8c6a42-d5b6-44bf-a2e4-747eb11670ce {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.757700] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b3fbed-4ca8-46f1-95d8-be72977ab778 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.761520] env[63293]: DEBUG oslo_vmware.api [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for the task: (returnval){ [ 849.761520] env[63293]: value = "task-1327681" [ 849.761520] env[63293]: _type = "Task" [ 849.761520] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.773062] env[63293]: DEBUG nova.compute.provider_tree [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.784577] env[63293]: DEBUG oslo_vmware.api [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327681, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.789771] env[63293]: DEBUG oslo_vmware.api [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327677, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.223365} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.790020] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.790243] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 849.790432] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 849.790619] env[63293]: INFO nova.compute.manager [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Took 1.14 seconds to destroy the instance on the hypervisor. [ 849.790868] env[63293]: DEBUG oslo.service.loopingcall [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.791068] env[63293]: DEBUG nova.compute.manager [-] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.791166] env[63293]: DEBUG nova.network.neutron [-] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 849.944371] env[63293]: DEBUG nova.network.neutron [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updated VIF entry in instance network info cache for port 26b86f34-f9b6-4b0b-abdd-d8acd51085fd. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 849.944755] env[63293]: DEBUG nova.network.neutron [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updating instance_info_cache with network_info: [{"id": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "address": "fa:16:3e:fc:c1:7d", "network": {"id": "bbf31454-8e46-45d9-b10a-fccda17654c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1875587777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "6fa6cd8dea16412db105faf2ff93a6ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26b86f34-f9", "ovs_interfaceid": "26b86f34-f9b6-4b0b-abdd-d8acd51085fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.014451] env[63293]: DEBUG oslo_vmware.api [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Task: {'id': task-1327679, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.359874} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.014676] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.014862] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 850.015356] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 850.015614] env[63293]: INFO nova.compute.manager [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Took 1.27 seconds to destroy the instance on the hypervisor. [ 850.015899] env[63293]: DEBUG oslo.service.loopingcall [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.016069] env[63293]: DEBUG nova.compute.manager [-] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.016170] env[63293]: DEBUG nova.network.neutron [-] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 850.022939] env[63293]: INFO nova.compute.manager [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Took 30.12 seconds to build instance. [ 850.076017] env[63293]: DEBUG nova.compute.manager [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 850.275861] env[63293]: DEBUG oslo_vmware.api [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Task: {'id': task-1327681, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369856} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.275861] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.275861] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 850.275861] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 850.275861] env[63293]: INFO nova.compute.manager [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Took 1.16 seconds to destroy the instance on the hypervisor. [ 850.276147] env[63293]: DEBUG oslo.service.loopingcall [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.276147] env[63293]: DEBUG nova.compute.manager [-] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.276147] env[63293]: DEBUG nova.network.neutron [-] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 850.285791] env[63293]: DEBUG nova.scheduler.client.report [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.321147] env[63293]: DEBUG nova.compute.manager [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.349371] env[63293]: DEBUG nova.virt.hardware [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.350435] env[63293]: DEBUG nova.virt.hardware [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.350435] env[63293]: DEBUG nova.virt.hardware [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.350435] env[63293]: DEBUG nova.virt.hardware [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.350435] env[63293]: DEBUG nova.virt.hardware [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.350435] env[63293]: DEBUG nova.virt.hardware [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.350645] env[63293]: DEBUG nova.virt.hardware [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.350686] env[63293]: DEBUG nova.virt.hardware [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.351074] env[63293]: DEBUG nova.virt.hardware [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.351074] env[63293]: DEBUG nova.virt.hardware [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.351173] env[63293]: DEBUG nova.virt.hardware [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.355586] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5868da-1f86-4a0c-8a40-222286307944 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.360164] env[63293]: DEBUG nova.compute.manager [req-ef0658dc-4c1d-4308-86bb-ca02edf822f2 req-87e03a83-656b-4c8a-ab21-0fc36e96944e service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Received event network-vif-deleted-26b86f34-f9b6-4b0b-abdd-d8acd51085fd {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.360481] env[63293]: INFO nova.compute.manager [req-ef0658dc-4c1d-4308-86bb-ca02edf822f2 req-87e03a83-656b-4c8a-ab21-0fc36e96944e service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Neutron deleted interface 26b86f34-f9b6-4b0b-abdd-d8acd51085fd; detaching it from the instance and deleting it from the info cache [ 850.360546] env[63293]: DEBUG nova.network.neutron [req-ef0658dc-4c1d-4308-86bb-ca02edf822f2 req-87e03a83-656b-4c8a-ab21-0fc36e96944e service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.369621] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b68fab-e866-4f9d-ada6-9a7c177c279c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.447836] env[63293]: DEBUG oslo_concurrency.lockutils [req-893afcc7-4843-42cc-ad31-74a77d852d37 req-d64168c1-448b-4ea3-bd08-e639395a27b3 service nova] Releasing lock "refresh_cache-48603812-c3cc-4dae-b8ba-99b9ac5f7969" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.523866] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f9ce89a-0113-4c51-ac25-40b533cb028c tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Lock "93080446-fd45-4281-bf6e-35d1d964f41e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.440s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.599940] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.607583] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquiring lock "93080446-fd45-4281-bf6e-35d1d964f41e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.607858] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Lock "93080446-fd45-4281-bf6e-35d1d964f41e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.608166] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquiring lock "93080446-fd45-4281-bf6e-35d1d964f41e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.608266] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Lock "93080446-fd45-4281-bf6e-35d1d964f41e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.608432] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Lock "93080446-fd45-4281-bf6e-35d1d964f41e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.611035] env[63293]: INFO nova.compute.manager [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Terminating instance [ 850.613338] env[63293]: DEBUG nova.compute.manager [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 850.613603] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 850.614763] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-814f5349-3ea5-4194-8e81-16becd3388a0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.625727] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.626400] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d2f4582-6382-4417-8d89-eada96f791f3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.636823] env[63293]: DEBUG oslo_vmware.api [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for the task: (returnval){ [ 850.636823] env[63293]: value = "task-1327682" [ 850.636823] env[63293]: _type = "Task" [ 850.636823] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.646020] env[63293]: DEBUG oslo_vmware.api [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327682, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.755240] env[63293]: DEBUG nova.network.neutron [-] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.793149] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.793994] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 850.797583] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.703s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.799484] env[63293]: INFO nova.compute.claims [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.802855] env[63293]: DEBUG nova.network.neutron [-] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.863669] env[63293]: DEBUG nova.network.neutron [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Successfully updated port: ee3d149b-d66e-439b-a49e-6f2e6c9d9495 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 850.867909] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-71d3ec2e-f196-4d77-b68e-3c44ace30f39 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.881996] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba25ce8-aad4-470b-b42e-73a8e6a36bad {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.911541] env[63293]: DEBUG nova.compute.manager [req-ef0658dc-4c1d-4308-86bb-ca02edf822f2 req-87e03a83-656b-4c8a-ab21-0fc36e96944e service nova] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Detach interface failed, port_id=26b86f34-f9b6-4b0b-abdd-d8acd51085fd, reason: Instance 48603812-c3cc-4dae-b8ba-99b9ac5f7969 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 850.998499] env[63293]: DEBUG nova.compute.manager [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Received event network-changed-b6279a11-248c-4e66-9df7-aa175821b8f1 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.998781] env[63293]: DEBUG nova.compute.manager [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Refreshing instance network info cache due to event network-changed-b6279a11-248c-4e66-9df7-aa175821b8f1. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.999063] env[63293]: DEBUG oslo_concurrency.lockutils [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] Acquiring lock "refresh_cache-f427933b-dd2c-49a9-b401-a647635a1eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.999280] env[63293]: DEBUG oslo_concurrency.lockutils [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] Acquired lock "refresh_cache-f427933b-dd2c-49a9-b401-a647635a1eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.999513] env[63293]: DEBUG nova.network.neutron [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Refreshing network info cache for port b6279a11-248c-4e66-9df7-aa175821b8f1 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.025410] env[63293]: DEBUG nova.network.neutron [-] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.028805] env[63293]: DEBUG nova.compute.manager [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 851.148203] env[63293]: DEBUG oslo_vmware.api [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327682, 'name': PowerOffVM_Task, 'duration_secs': 0.241744} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.148483] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 851.148654] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 851.148928] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82763dab-5cc8-4d0f-9fd2-f46b9f301833 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.242735] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 851.243108] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 851.243211] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Deleting the datastore file [datastore1] 93080446-fd45-4281-bf6e-35d1d964f41e {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.243464] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fe301f9-f914-4969-bed3-5b133c4d0742 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.252615] env[63293]: DEBUG oslo_vmware.api [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for the task: (returnval){ [ 851.252615] env[63293]: value = "task-1327684" [ 851.252615] env[63293]: _type = "Task" [ 851.252615] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.261137] env[63293]: INFO nova.compute.manager [-] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Took 1.47 seconds to deallocate network for instance. [ 851.261484] env[63293]: DEBUG oslo_vmware.api [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327684, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.307947] env[63293]: DEBUG nova.compute.utils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.309263] env[63293]: INFO nova.compute.manager [-] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Took 1.29 seconds to deallocate network for instance. [ 851.310083] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 851.310255] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 851.353402] env[63293]: DEBUG nova.policy [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c34e72ea18dd409ba5d40b32711f456f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49f167005da748fcaf2497a5040a838c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 851.367987] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.371114] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.371114] env[63293]: DEBUG nova.network.neutron [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.530214] env[63293]: INFO nova.compute.manager [-] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Took 1.25 seconds to deallocate network for instance. [ 851.544236] env[63293]: DEBUG nova.network.neutron [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.574311] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.659707] env[63293]: DEBUG nova.network.neutron [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.735687] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Successfully created port: c7cbdf4e-1a17-436f-8417-ed153aa1e71b {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 851.763088] env[63293]: DEBUG oslo_vmware.api [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Task: {'id': task-1327684, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189545} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.763451] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.763724] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 851.764356] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 851.764620] env[63293]: INFO nova.compute.manager [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 851.764946] env[63293]: DEBUG oslo.service.loopingcall [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.765218] env[63293]: DEBUG nova.compute.manager [-] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 851.765573] env[63293]: DEBUG nova.network.neutron [-] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 851.769196] env[63293]: DEBUG oslo_concurrency.lockutils [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.814681] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 851.828151] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.922362] env[63293]: DEBUG nova.network.neutron [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.039863] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.126369] env[63293]: DEBUG nova.network.neutron [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Updating instance_info_cache with network_info: [{"id": "ee3d149b-d66e-439b-a49e-6f2e6c9d9495", "address": "fa:16:3e:30:5d:6f", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee3d149b-d6", "ovs_interfaceid": "ee3d149b-d66e-439b-a49e-6f2e6c9d9495", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.163023] env[63293]: DEBUG oslo_concurrency.lockutils [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] Releasing lock "refresh_cache-f427933b-dd2c-49a9-b401-a647635a1eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.163147] env[63293]: DEBUG nova.compute.manager [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Received event network-changed-4dbccf6a-053e-4933-91bc-1b7af5a260be {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.163335] env[63293]: DEBUG nova.compute.manager [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Refreshing instance network info cache due to event network-changed-4dbccf6a-053e-4933-91bc-1b7af5a260be. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 852.163717] env[63293]: DEBUG oslo_concurrency.lockutils [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] Acquiring lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.163717] env[63293]: DEBUG oslo_concurrency.lockutils [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] Acquired lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.163819] env[63293]: DEBUG nova.network.neutron [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Refreshing network info cache for port 4dbccf6a-053e-4933-91bc-1b7af5a260be {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 852.184306] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86ab88f-3c4b-49f6-b5e2-bb5a6cc5f465 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.196401] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bbc9104-034c-4eec-844f-07e083a53441 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.232677] env[63293]: DEBUG nova.network.neutron [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Successfully updated port: 31fd0cd3-702f-4cea-860d-fc21f02effb2 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.234320] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a18f1d-0c96-40fe-a2af-371bab2bab21 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.242911] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-450917cf-6dda-4771-b6d1-781eaf1332e1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.257068] env[63293]: DEBUG nova.compute.provider_tree [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.387450] env[63293]: DEBUG nova.compute.manager [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Received event network-vif-plugged-ee3d149b-d66e-439b-a49e-6f2e6c9d9495 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.387682] env[63293]: DEBUG oslo_concurrency.lockutils [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] Acquiring lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.387876] env[63293]: DEBUG oslo_concurrency.lockutils [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.388085] env[63293]: DEBUG oslo_concurrency.lockutils [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.388276] env[63293]: DEBUG nova.compute.manager [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] No waiting events found dispatching network-vif-plugged-ee3d149b-d66e-439b-a49e-6f2e6c9d9495 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.388489] env[63293]: WARNING nova.compute.manager [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Received unexpected event network-vif-plugged-ee3d149b-d66e-439b-a49e-6f2e6c9d9495 for instance with vm_state building and task_state spawning. [ 852.388716] env[63293]: DEBUG nova.compute.manager [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Received event network-changed-ee3d149b-d66e-439b-a49e-6f2e6c9d9495 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.388892] env[63293]: DEBUG nova.compute.manager [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Refreshing instance network info cache due to event network-changed-ee3d149b-d66e-439b-a49e-6f2e6c9d9495. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 852.389143] env[63293]: DEBUG oslo_concurrency.lockutils [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] Acquiring lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.629183] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.629537] env[63293]: DEBUG nova.compute.manager [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Instance network_info: |[{"id": "ee3d149b-d66e-439b-a49e-6f2e6c9d9495", "address": "fa:16:3e:30:5d:6f", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee3d149b-d6", "ovs_interfaceid": "ee3d149b-d66e-439b-a49e-6f2e6c9d9495", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 852.630286] env[63293]: DEBUG oslo_concurrency.lockutils [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] Acquired lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.630634] env[63293]: DEBUG nova.network.neutron [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Refreshing network info cache for port ee3d149b-d66e-439b-a49e-6f2e6c9d9495 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 852.632108] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:5d:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee3d149b-d66e-439b-a49e-6f2e6c9d9495', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 852.639822] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Creating folder: Project (ff8479fc251544e9b4618deccf992754). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.640528] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2f29e8ea-7391-4054-b628-a8b9c5335f10 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.652834] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Created folder: Project (ff8479fc251544e9b4618deccf992754) in parent group-v283678. [ 852.653066] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Creating folder: Instances. Parent ref: group-v283752. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.654115] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2b0ce7ac-c38d-4a8d-abd3-08552be9687c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.655031] env[63293]: DEBUG nova.network.neutron [-] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.662897] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Created folder: Instances in parent group-v283752. [ 852.663167] env[63293]: DEBUG oslo.service.loopingcall [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.663376] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 852.663592] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e709843-d0af-444b-a377-a16a6b74fd2a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.686980] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.686980] env[63293]: value = "task-1327687" [ 852.686980] env[63293]: _type = "Task" [ 852.686980] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.695774] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327687, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.740224] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "refresh_cache-f427933b-dd2c-49a9-b401-a647635a1eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.740224] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquired lock "refresh_cache-f427933b-dd2c-49a9-b401-a647635a1eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.740224] env[63293]: DEBUG nova.network.neutron [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 852.761346] env[63293]: DEBUG nova.scheduler.client.report [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.837024] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 852.866927] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 852.867226] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 852.867390] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.867574] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 852.867767] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.867862] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 852.868106] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 852.868316] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 852.868497] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 852.868664] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 852.868834] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.869738] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7cd552-2f7c-4d83-9593-3d3ebd0d06af {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.879909] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9059374-b341-47a1-8902-424cbb9594ca {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.993718] env[63293]: DEBUG nova.network.neutron [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updated VIF entry in instance network info cache for port 4dbccf6a-053e-4933-91bc-1b7af5a260be. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 852.994125] env[63293]: DEBUG nova.network.neutron [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance_info_cache with network_info: [{"id": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "address": "fa:16:3e:93:d3:25", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dbccf6a-05", "ovs_interfaceid": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.024718] env[63293]: DEBUG nova.compute.manager [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Received event network-vif-plugged-31fd0cd3-702f-4cea-860d-fc21f02effb2 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.024936] env[63293]: DEBUG oslo_concurrency.lockutils [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] Acquiring lock "f427933b-dd2c-49a9-b401-a647635a1eab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.025209] env[63293]: DEBUG oslo_concurrency.lockutils [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] Lock "f427933b-dd2c-49a9-b401-a647635a1eab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.025368] env[63293]: DEBUG oslo_concurrency.lockutils [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] Lock "f427933b-dd2c-49a9-b401-a647635a1eab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.025525] env[63293]: DEBUG nova.compute.manager [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] No waiting events found dispatching network-vif-plugged-31fd0cd3-702f-4cea-860d-fc21f02effb2 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 853.025693] env[63293]: WARNING nova.compute.manager [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Received unexpected event network-vif-plugged-31fd0cd3-702f-4cea-860d-fc21f02effb2 for instance with vm_state building and task_state spawning. [ 853.025859] env[63293]: DEBUG nova.compute.manager [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Received event network-vif-deleted-20da136a-9449-4b46-be97-b9be3c4c2bfe {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.026042] env[63293]: DEBUG nova.compute.manager [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Received event network-changed-31fd0cd3-702f-4cea-860d-fc21f02effb2 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.026207] env[63293]: DEBUG nova.compute.manager [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Refreshing instance network info cache due to event network-changed-31fd0cd3-702f-4cea-860d-fc21f02effb2. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 853.026379] env[63293]: DEBUG oslo_concurrency.lockutils [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] Acquiring lock "refresh_cache-f427933b-dd2c-49a9-b401-a647635a1eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.160997] env[63293]: INFO nova.compute.manager [-] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Took 1.40 seconds to deallocate network for instance. [ 853.197574] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327687, 'name': CreateVM_Task, 'duration_secs': 0.35617} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.198420] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 853.199126] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.199306] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.199674] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 853.200192] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-493ec023-561d-48e8-8bf3-d5698cc033ec {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.208032] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 853.208032] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5268b444-ceae-63bb-2234-b5e2b884a522" [ 853.208032] env[63293]: _type = "Task" [ 853.208032] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.216056] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5268b444-ceae-63bb-2234-b5e2b884a522, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.269303] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.270050] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 853.272532] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.442s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.272751] env[63293]: DEBUG nova.objects.instance [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lazy-loading 'resources' on Instance uuid cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.347953] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Successfully updated port: c7cbdf4e-1a17-436f-8417-ed153aa1e71b {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 853.478833] env[63293]: DEBUG nova.network.neutron [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 853.496871] env[63293]: DEBUG oslo_concurrency.lockutils [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] Releasing lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.497210] env[63293]: DEBUG nova.compute.manager [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Received event network-vif-deleted-8b8079ba-2afe-42c9-9536-3b8951b51e00 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.497404] env[63293]: DEBUG nova.compute.manager [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Received event network-vif-deleted-f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.497647] env[63293]: INFO nova.compute.manager [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Neutron deleted interface f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13; detaching it from the instance and deleting it from the info cache [ 853.497845] env[63293]: DEBUG nova.network.neutron [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.649339] env[63293]: DEBUG nova.network.neutron [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Updated VIF entry in instance network info cache for port ee3d149b-d66e-439b-a49e-6f2e6c9d9495. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 853.649843] env[63293]: DEBUG nova.network.neutron [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Updating instance_info_cache with network_info: [{"id": "ee3d149b-d66e-439b-a49e-6f2e6c9d9495", "address": "fa:16:3e:30:5d:6f", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee3d149b-d6", "ovs_interfaceid": "ee3d149b-d66e-439b-a49e-6f2e6c9d9495", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.667741] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.722376] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5268b444-ceae-63bb-2234-b5e2b884a522, 'name': SearchDatastore_Task, 'duration_secs': 0.011224} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.722702] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.722934] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.723178] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.723331] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.723512] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.723777] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8bb80028-38a4-4de1-a753-3f1f6effefd8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.732541] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.732687] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.733439] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bf175e8-eb00-4509-a2c6-9b33eeaa15e5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.740742] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 853.740742] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a1d0d9-cfb2-146d-9793-5e5938465cc3" [ 853.740742] env[63293]: _type = "Task" [ 853.740742] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.748198] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a1d0d9-cfb2-146d-9793-5e5938465cc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.776188] env[63293]: DEBUG nova.compute.utils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.779985] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 853.780191] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 853.849576] env[63293]: DEBUG nova.policy [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c34e72ea18dd409ba5d40b32711f456f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49f167005da748fcaf2497a5040a838c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 853.853636] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "refresh_cache-8ea7f3c3-8d73-4fb2-a848-4be4649e5660" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.853788] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired lock "refresh_cache-8ea7f3c3-8d73-4fb2-a848-4be4649e5660" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.853915] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.922525] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.922905] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.932882] env[63293]: DEBUG nova.network.neutron [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Updating instance_info_cache with network_info: [{"id": "b6279a11-248c-4e66-9df7-aa175821b8f1", "address": "fa:16:3e:2b:be:4b", "network": {"id": "215cf030-8fa2-4372-a67f-219d6568ade3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1068518024", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d72964783a4d0e95daa2305c6148cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6279a11-24", "ovs_interfaceid": "b6279a11-248c-4e66-9df7-aa175821b8f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "31fd0cd3-702f-4cea-860d-fc21f02effb2", "address": "fa:16:3e:e3:88:d6", "network": {"id": "14998d91-038c-4784-9973-9315d11294eb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1366276912", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.28", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "09d72964783a4d0e95daa2305c6148cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31fd0cd3-70", "ovs_interfaceid": "31fd0cd3-702f-4cea-860d-fc21f02effb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.002989] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-40145ce8-2700-48e3-b9b3-72ee0cf598cc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.013022] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab0ec8f-deee-4367-8fea-764559377f32 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.042764] env[63293]: DEBUG nova.compute.manager [req-2687e44f-7fa3-47c3-8e8a-142ef24b7565 req-1ef8e8f2-7555-4d0f-85eb-2579e22203a3 service nova] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Detach interface failed, port_id=f8b1091f-58a0-4f84-aa66-8c0e4b8b6b13, reason: Instance 4593cabd-1331-4ee1-99f5-46edeac56859 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 854.146605] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5c1d20-a2d2-40ed-8108-4eb7eb5717cb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.153397] env[63293]: DEBUG oslo_concurrency.lockutils [req-31007d7e-992c-4d25-910c-01f1604220e7 req-bce4fd98-c748-4bc0-8e12-f95c2bc36b1a service nova] Releasing lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.154531] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eff9c2a-53eb-4358-b6f5-68c2437baaac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.183628] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88007bd8-6eef-4046-bd52-1cdb1b6857e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.191496] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546e2066-174c-4d77-b011-7acf8926564e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.206459] env[63293]: DEBUG nova.compute.provider_tree [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.208357] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Successfully created port: 53a4ce4d-f5b1-46ed-b4fe-db9454cef578 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.250719] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a1d0d9-cfb2-146d-9793-5e5938465cc3, 'name': SearchDatastore_Task, 'duration_secs': 0.009004} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.251511] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f35321a4-2acb-47d5-82e1-52b87cb5bc41 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.256563] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 854.256563] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52eeb9a4-12d2-9a6b-e664-8ab1e36b859a" [ 854.256563] env[63293]: _type = "Task" [ 854.256563] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.263706] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52eeb9a4-12d2-9a6b-e664-8ab1e36b859a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.280434] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 854.397119] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.411174] env[63293]: DEBUG nova.compute.manager [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Received event network-vif-plugged-c7cbdf4e-1a17-436f-8417-ed153aa1e71b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.411437] env[63293]: DEBUG oslo_concurrency.lockutils [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] Acquiring lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.411591] env[63293]: DEBUG oslo_concurrency.lockutils [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] Lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.411795] env[63293]: DEBUG oslo_concurrency.lockutils [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] Lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.411967] env[63293]: DEBUG nova.compute.manager [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] No waiting events found dispatching network-vif-plugged-c7cbdf4e-1a17-436f-8417-ed153aa1e71b {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.412194] env[63293]: WARNING nova.compute.manager [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Received unexpected event network-vif-plugged-c7cbdf4e-1a17-436f-8417-ed153aa1e71b for instance with vm_state building and task_state spawning. [ 854.412347] env[63293]: DEBUG nova.compute.manager [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Received event network-changed-c7cbdf4e-1a17-436f-8417-ed153aa1e71b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.412509] env[63293]: DEBUG nova.compute.manager [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Refreshing instance network info cache due to event network-changed-c7cbdf4e-1a17-436f-8417-ed153aa1e71b. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.412680] env[63293]: DEBUG oslo_concurrency.lockutils [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] Acquiring lock "refresh_cache-8ea7f3c3-8d73-4fb2-a848-4be4649e5660" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.427440] env[63293]: DEBUG nova.compute.utils [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.434822] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Releasing lock "refresh_cache-f427933b-dd2c-49a9-b401-a647635a1eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.435296] env[63293]: DEBUG nova.compute.manager [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Instance network_info: |[{"id": "b6279a11-248c-4e66-9df7-aa175821b8f1", "address": "fa:16:3e:2b:be:4b", "network": {"id": "215cf030-8fa2-4372-a67f-219d6568ade3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1068518024", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d72964783a4d0e95daa2305c6148cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6279a11-24", "ovs_interfaceid": "b6279a11-248c-4e66-9df7-aa175821b8f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "31fd0cd3-702f-4cea-860d-fc21f02effb2", "address": "fa:16:3e:e3:88:d6", "network": {"id": "14998d91-038c-4784-9973-9315d11294eb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1366276912", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.28", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "09d72964783a4d0e95daa2305c6148cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31fd0cd3-70", "ovs_interfaceid": "31fd0cd3-702f-4cea-860d-fc21f02effb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 854.435296] env[63293]: DEBUG oslo_concurrency.lockutils [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] Acquired lock "refresh_cache-f427933b-dd2c-49a9-b401-a647635a1eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.435649] env[63293]: DEBUG nova.network.neutron [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Refreshing network info cache for port 31fd0cd3-702f-4cea-860d-fc21f02effb2 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.436730] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:be:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6279a11-248c-4e66-9df7-aa175821b8f1', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:88:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9297313e-7c50-4873-93d3-67284929163a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31fd0cd3-702f-4cea-860d-fc21f02effb2', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.447323] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Creating folder: Project (09d72964783a4d0e95daa2305c6148cb). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.452027] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af50d94a-16f5-466b-923e-5fadf972dedb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.461251] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Created folder: Project (09d72964783a4d0e95daa2305c6148cb) in parent group-v283678. [ 854.461473] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Creating folder: Instances. Parent ref: group-v283755. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.461739] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-61adc34c-50cc-41f2-a27f-b82c5e304e9c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.474800] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Created folder: Instances in parent group-v283755. [ 854.475052] env[63293]: DEBUG oslo.service.loopingcall [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.475244] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.475451] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e2e73a4-2469-44af-bc69-b04c6de4aaff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.496723] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.496723] env[63293]: value = "task-1327690" [ 854.496723] env[63293]: _type = "Task" [ 854.496723] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.506477] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327690, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.592736] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Updating instance_info_cache with network_info: [{"id": "c7cbdf4e-1a17-436f-8417-ed153aa1e71b", "address": "fa:16:3e:aa:c8:d4", "network": {"id": "b69cd9c0-f181-4f90-ac61-a0ebe2ea70b5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1584936849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49f167005da748fcaf2497a5040a838c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7cbdf4e-1a", "ovs_interfaceid": "c7cbdf4e-1a17-436f-8417-ed153aa1e71b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.711485] env[63293]: DEBUG nova.scheduler.client.report [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.767143] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52eeb9a4-12d2-9a6b-e664-8ab1e36b859a, 'name': SearchDatastore_Task, 'duration_secs': 0.009328} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.767568] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.767764] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 6cd8b0eb-4bfe-49e0-b392-900f50865ad9/6cd8b0eb-4bfe-49e0-b392-900f50865ad9.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 854.768031] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-019282a2-8620-4268-b68b-fe202d584af7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.776135] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 854.776135] env[63293]: value = "task-1327691" [ 854.776135] env[63293]: _type = "Task" [ 854.776135] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.788726] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327691, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.930199] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.009267] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327690, 'name': CreateVM_Task, 'duration_secs': 0.506201} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.009267] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.010073] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.010517] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.010688] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.010987] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c69e4b6-26d4-4309-acc5-a86d565d51a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.015960] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for the task: (returnval){ [ 855.015960] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52606f94-39ed-96e7-e51a-ea8602674728" [ 855.015960] env[63293]: _type = "Task" [ 855.015960] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.024477] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52606f94-39ed-96e7-e51a-ea8602674728, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.094926] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Releasing lock "refresh_cache-8ea7f3c3-8d73-4fb2-a848-4be4649e5660" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.095283] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Instance network_info: |[{"id": "c7cbdf4e-1a17-436f-8417-ed153aa1e71b", "address": "fa:16:3e:aa:c8:d4", "network": {"id": "b69cd9c0-f181-4f90-ac61-a0ebe2ea70b5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1584936849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49f167005da748fcaf2497a5040a838c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7cbdf4e-1a", "ovs_interfaceid": "c7cbdf4e-1a17-436f-8417-ed153aa1e71b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.095650] env[63293]: DEBUG oslo_concurrency.lockutils [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] Acquired lock "refresh_cache-8ea7f3c3-8d73-4fb2-a848-4be4649e5660" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.095845] env[63293]: DEBUG nova.network.neutron [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Refreshing network info cache for port c7cbdf4e-1a17-436f-8417-ed153aa1e71b {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 855.097087] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:c8:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35e463c7-7d78-4d66-8efd-6127b1f3ee17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7cbdf4e-1a17-436f-8417-ed153aa1e71b', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.107308] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Creating folder: Project (49f167005da748fcaf2497a5040a838c). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.108527] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bf675eb5-58ff-4e2f-90b7-cee82d29cc04 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.120059] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Created folder: Project (49f167005da748fcaf2497a5040a838c) in parent group-v283678. [ 855.120609] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Creating folder: Instances. Parent ref: group-v283758. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.120609] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-77949b05-d621-4315-bd9f-ab72051e6692 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.131651] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Created folder: Instances in parent group-v283758. [ 855.131899] env[63293]: DEBUG oslo.service.loopingcall [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.132110] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 855.132326] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7525148f-45f3-4946-8e18-1a342d7c46ab {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.155716] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.155716] env[63293]: value = "task-1327694" [ 855.155716] env[63293]: _type = "Task" [ 855.155716] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.163824] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327694, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.217765] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.945s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.221164] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.085s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.221391] env[63293]: DEBUG nova.objects.instance [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63293) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 855.250429] env[63293]: INFO nova.scheduler.client.report [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleted allocations for instance cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395 [ 855.265705] env[63293]: DEBUG nova.network.neutron [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Updated VIF entry in instance network info cache for port 31fd0cd3-702f-4cea-860d-fc21f02effb2. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 855.266643] env[63293]: DEBUG nova.network.neutron [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Updating instance_info_cache with network_info: [{"id": "b6279a11-248c-4e66-9df7-aa175821b8f1", "address": "fa:16:3e:2b:be:4b", "network": {"id": "215cf030-8fa2-4372-a67f-219d6568ade3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1068518024", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d72964783a4d0e95daa2305c6148cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6279a11-24", "ovs_interfaceid": "b6279a11-248c-4e66-9df7-aa175821b8f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "31fd0cd3-702f-4cea-860d-fc21f02effb2", "address": "fa:16:3e:e3:88:d6", "network": {"id": "14998d91-038c-4784-9973-9315d11294eb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1366276912", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.28", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "09d72964783a4d0e95daa2305c6148cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9297313e-7c50-4873-93d3-67284929163a", "external-id": "nsx-vlan-transportzone-620", "segmentation_id": 620, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31fd0cd3-70", "ovs_interfaceid": "31fd0cd3-702f-4cea-860d-fc21f02effb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.286687] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327691, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484583} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.286961] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 6cd8b0eb-4bfe-49e0-b392-900f50865ad9/6cd8b0eb-4bfe-49e0-b392-900f50865ad9.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 855.287273] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 855.287477] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9766a928-1966-41f2-968a-71bf3c3e9de7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.292816] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 855.302927] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 855.302927] env[63293]: value = "task-1327695" [ 855.302927] env[63293]: _type = "Task" [ 855.302927] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.311454] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327695, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.328797] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.328797] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.328953] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.329160] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.329336] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.329502] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.329732] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.329895] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.330088] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.330271] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.330457] env[63293]: DEBUG nova.virt.hardware [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.331353] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b41762c-22e2-4644-91f7-963f6fe7a96e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.342168] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7fd25b-cf33-4c03-85e7-632c0abb6ac6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.380765] env[63293]: DEBUG nova.network.neutron [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Updated VIF entry in instance network info cache for port c7cbdf4e-1a17-436f-8417-ed153aa1e71b. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 855.381174] env[63293]: DEBUG nova.network.neutron [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Updating instance_info_cache with network_info: [{"id": "c7cbdf4e-1a17-436f-8417-ed153aa1e71b", "address": "fa:16:3e:aa:c8:d4", "network": {"id": "b69cd9c0-f181-4f90-ac61-a0ebe2ea70b5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1584936849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49f167005da748fcaf2497a5040a838c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7cbdf4e-1a", "ovs_interfaceid": "c7cbdf4e-1a17-436f-8417-ed153aa1e71b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.526499] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52606f94-39ed-96e7-e51a-ea8602674728, 'name': SearchDatastore_Task, 'duration_secs': 0.018042} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.526820] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.527095] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.527366] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.527523] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.527710] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.527975] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c3a9d7a8-15ff-4269-a75a-ed4cb8ca6c56 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.537737] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.538289] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.538620] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c954ad2-00f7-42e8-b894-5cf5efd6592d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.543464] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for the task: (returnval){ [ 855.543464] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5266fe88-c280-fb70-640e-f738f3095b31" [ 855.543464] env[63293]: _type = "Task" [ 855.543464] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.550945] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5266fe88-c280-fb70-640e-f738f3095b31, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.667245] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327694, 'name': CreateVM_Task, 'duration_secs': 0.325878} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.667245] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.667245] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.667245] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.667245] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.667245] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9383aa7c-5c7b-4243-a88b-13d5f7186b3b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.671527] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 855.671527] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b7e36f-d1d0-e55c-405c-69b9ecb47710" [ 855.671527] env[63293]: _type = "Task" [ 855.671527] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.679817] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b7e36f-d1d0-e55c-405c-69b9ecb47710, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.704316] env[63293]: DEBUG nova.compute.manager [req-2d56c303-5b5d-4773-8179-594fac5fb8a4 req-085c9401-4216-4df8-87ec-a5820c39b147 service nova] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Received event network-vif-plugged-53a4ce4d-f5b1-46ed-b4fe-db9454cef578 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.704645] env[63293]: DEBUG oslo_concurrency.lockutils [req-2d56c303-5b5d-4773-8179-594fac5fb8a4 req-085c9401-4216-4df8-87ec-a5820c39b147 service nova] Acquiring lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.704935] env[63293]: DEBUG oslo_concurrency.lockutils [req-2d56c303-5b5d-4773-8179-594fac5fb8a4 req-085c9401-4216-4df8-87ec-a5820c39b147 service nova] Lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.705180] env[63293]: DEBUG oslo_concurrency.lockutils [req-2d56c303-5b5d-4773-8179-594fac5fb8a4 req-085c9401-4216-4df8-87ec-a5820c39b147 service nova] Lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.705411] env[63293]: DEBUG nova.compute.manager [req-2d56c303-5b5d-4773-8179-594fac5fb8a4 req-085c9401-4216-4df8-87ec-a5820c39b147 service nova] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] No waiting events found dispatching network-vif-plugged-53a4ce4d-f5b1-46ed-b4fe-db9454cef578 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.705643] env[63293]: WARNING nova.compute.manager [req-2d56c303-5b5d-4773-8179-594fac5fb8a4 req-085c9401-4216-4df8-87ec-a5820c39b147 service nova] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Received unexpected event network-vif-plugged-53a4ce4d-f5b1-46ed-b4fe-db9454cef578 for instance with vm_state building and task_state spawning. [ 855.761601] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa497dd5-ec84-4e24-a225-3fc054daa67d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.927s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.768424] env[63293]: DEBUG oslo_concurrency.lockutils [req-91432187-15c0-4124-84fa-476f5708d9ae req-942e0167-759d-4d0c-8774-e3fcd7e0e55b service nova] Releasing lock "refresh_cache-f427933b-dd2c-49a9-b401-a647635a1eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.814185] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327695, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063907} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.814468] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.815264] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e065187d-dce7-4225-a476-6912b3bcce8a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.837560] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 6cd8b0eb-4bfe-49e0-b392-900f50865ad9/6cd8b0eb-4bfe-49e0-b392-900f50865ad9.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.838524] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Successfully updated port: 53a4ce4d-f5b1-46ed-b4fe-db9454cef578 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 855.840096] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58aa0534-23e0-43cd-a7d0-ed022e760b0e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.856031] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "refresh_cache-9973dd7f-ca75-4db0-8d2b-f21b0311abeb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.856185] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired lock "refresh_cache-9973dd7f-ca75-4db0-8d2b-f21b0311abeb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.856357] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 855.862693] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 855.862693] env[63293]: value = "task-1327696" [ 855.862693] env[63293]: _type = "Task" [ 855.862693] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.871623] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327696, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.883728] env[63293]: DEBUG oslo_concurrency.lockutils [req-7a10ee5c-e828-4051-988a-67ea78766a16 req-f2debb64-562a-4b41-8599-684ef91308d1 service nova] Releasing lock "refresh_cache-8ea7f3c3-8d73-4fb2-a848-4be4649e5660" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.992954] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.992954] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.992954] env[63293]: INFO nova.compute.manager [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Attaching volume 981997fd-c500-4d22-bd91-77b5fb478502 to /dev/sdb [ 856.027860] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4503cc89-026e-48d6-b953-b3308a037b93 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.035111] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8e4234-4118-49a5-ad67-404e588ee70b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.051028] env[63293]: DEBUG nova.virt.block_device [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updating existing volume attachment record: d54ea2d2-47a7-4b8d-b3c4-d15910d0ee10 {{(pid=63293) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 856.058871] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5266fe88-c280-fb70-640e-f738f3095b31, 'name': SearchDatastore_Task, 'duration_secs': 0.007774} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.059688] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-082c89c9-4c33-4afc-bec8-21a8842a48d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.064881] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for the task: (returnval){ [ 856.064881] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]526ce704-fb67-3c02-33eb-290ea7064f3d" [ 856.064881] env[63293]: _type = "Task" [ 856.064881] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.075430] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]526ce704-fb67-3c02-33eb-290ea7064f3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.182088] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b7e36f-d1d0-e55c-405c-69b9ecb47710, 'name': SearchDatastore_Task, 'duration_secs': 0.008878} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.182379] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.182630] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.182890] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.183052] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.183567] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.183567] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be082a6f-606d-4c49-b376-3ab7cc2c5047 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.193636] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.193636] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 856.194187] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c64d328d-1475-4d0d-98b5-f2fa39027377 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.200096] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 856.200096] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5207939b-9181-7bd4-5c50-72d0a8b61be9" [ 856.200096] env[63293]: _type = "Task" [ 856.200096] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.207908] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5207939b-9181-7bd4-5c50-72d0a8b61be9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.233797] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a99aaa14-1f25-49a5-a50e-ae867032b3d1 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.234996] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.517s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.236620] env[63293]: INFO nova.compute.claims [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.362035] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 856.362035] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 856.373763] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327696, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.398061] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.550598] env[63293]: DEBUG nova.network.neutron [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Updating instance_info_cache with network_info: [{"id": "53a4ce4d-f5b1-46ed-b4fe-db9454cef578", "address": "fa:16:3e:8b:8f:9f", "network": {"id": "b69cd9c0-f181-4f90-ac61-a0ebe2ea70b5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1584936849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49f167005da748fcaf2497a5040a838c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53a4ce4d-f5", "ovs_interfaceid": "53a4ce4d-f5b1-46ed-b4fe-db9454cef578", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.574711] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]526ce704-fb67-3c02-33eb-290ea7064f3d, 'name': SearchDatastore_Task, 'duration_secs': 0.009698} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.574963] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.575243] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] f427933b-dd2c-49a9-b401-a647635a1eab/f427933b-dd2c-49a9-b401-a647635a1eab.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.575518] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2258236-b5a9-435f-8b55-773058359f93 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.581691] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for the task: (returnval){ [ 856.581691] env[63293]: value = "task-1327700" [ 856.581691] env[63293]: _type = "Task" [ 856.581691] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.590282] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327700, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.710889] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5207939b-9181-7bd4-5c50-72d0a8b61be9, 'name': SearchDatastore_Task, 'duration_secs': 0.008199} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.711639] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aabd5f3b-8cc2-4dbc-8543-d92e9b6c4ad9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.716802] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 856.716802] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fe27ae-5414-b044-1ad4-802c30787889" [ 856.716802] env[63293]: _type = "Task" [ 856.716802] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.724435] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fe27ae-5414-b044-1ad4-802c30787889, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.872826] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 856.873046] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Starting heal instance info cache {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 856.873126] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Rebuilding the list of instances to heal {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 856.883976] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327696, 'name': ReconfigVM_Task, 'duration_secs': 0.716927} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.884290] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 6cd8b0eb-4bfe-49e0-b392-900f50865ad9/6cd8b0eb-4bfe-49e0-b392-900f50865ad9.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.884969] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d0550ead-e17f-4979-b1e6-975cb3d2d8af {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.894024] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 856.894024] env[63293]: value = "task-1327701" [ 856.894024] env[63293]: _type = "Task" [ 856.894024] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.901986] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327701, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.053767] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Releasing lock "refresh_cache-9973dd7f-ca75-4db0-8d2b-f21b0311abeb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.054177] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Instance network_info: |[{"id": "53a4ce4d-f5b1-46ed-b4fe-db9454cef578", "address": "fa:16:3e:8b:8f:9f", "network": {"id": "b69cd9c0-f181-4f90-ac61-a0ebe2ea70b5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1584936849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49f167005da748fcaf2497a5040a838c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53a4ce4d-f5", "ovs_interfaceid": "53a4ce4d-f5b1-46ed-b4fe-db9454cef578", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 857.054599] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:8f:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35e463c7-7d78-4d66-8efd-6127b1f3ee17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '53a4ce4d-f5b1-46ed-b4fe-db9454cef578', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.062374] env[63293]: DEBUG oslo.service.loopingcall [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.062710] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 857.063144] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d06450b-4afe-440b-88be-e520ae45034e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.082921] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 857.082921] env[63293]: value = "task-1327702" [ 857.082921] env[63293]: _type = "Task" [ 857.082921] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.094206] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327700, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497323} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.096891] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] f427933b-dd2c-49a9-b401-a647635a1eab/f427933b-dd2c-49a9-b401-a647635a1eab.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 857.097121] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.097322] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327702, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.097513] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b466613-1d06-4c85-af7b-d5f712fcfd6a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.102917] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for the task: (returnval){ [ 857.102917] env[63293]: value = "task-1327703" [ 857.102917] env[63293]: _type = "Task" [ 857.102917] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.110678] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327703, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.227476] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fe27ae-5414-b044-1ad4-802c30787889, 'name': SearchDatastore_Task, 'duration_secs': 0.009281} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.227779] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.228047] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 8ea7f3c3-8d73-4fb2-a848-4be4649e5660/8ea7f3c3-8d73-4fb2-a848-4be4649e5660.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 857.228329] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6086658-aa1d-4f92-8740-1b81b2676524 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.234635] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 857.234635] env[63293]: value = "task-1327704" [ 857.234635] env[63293]: _type = "Task" [ 857.234635] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.242615] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327704, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.380115] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 857.380115] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 857.380249] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 857.380352] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 857.380518] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 857.405742] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327701, 'name': Rename_Task, 'duration_secs': 0.180931} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.406383] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 857.406752] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d930390-a219-4a69-a576-fd39a94ecbbc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.409528] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.409820] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquired lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.410066] env[63293]: DEBUG nova.network.neutron [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Forcefully refreshing network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 857.410332] env[63293]: DEBUG nova.objects.instance [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lazy-loading 'info_cache' on Instance uuid 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.419581] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 857.419581] env[63293]: value = "task-1327705" [ 857.419581] env[63293]: _type = "Task" [ 857.419581] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.431676] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327705, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.582076] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7586f0f4-650d-4698-a517-6880fc6bba6e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.593873] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74b242a-8f07-410a-81ab-87eeb4bd9073 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.600534] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327702, 'name': CreateVM_Task, 'duration_secs': 0.366778} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.601188] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 857.601981] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.602171] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.602568] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 857.603392] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36aff43d-d1a2-401d-bd27-013bdbc89ecd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.636299] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02055feb-2540-4b5c-b1ab-6427e6e5081a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.640536] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 857.640536] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f3d281-d36c-e1f7-9200-e93da86b7f2d" [ 857.640536] env[63293]: _type = "Task" [ 857.640536] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.648985] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327703, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071108} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.649794] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.651152] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3967184-392a-4a6a-b058-3710893320a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.658748] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce814683-b57d-47ac-bec2-f3b0c5360bd5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.661588] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f3d281-d36c-e1f7-9200-e93da86b7f2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.672789] env[63293]: DEBUG nova.compute.provider_tree [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.693873] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] f427933b-dd2c-49a9-b401-a647635a1eab/f427933b-dd2c-49a9-b401-a647635a1eab.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.695164] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d54a3902-eb31-4940-b10e-c42a0957c5c4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.715951] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for the task: (returnval){ [ 857.715951] env[63293]: value = "task-1327706" [ 857.715951] env[63293]: _type = "Task" [ 857.715951] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.726170] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327706, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.729084] env[63293]: DEBUG nova.compute.manager [req-092d11d0-5caa-4e36-840b-4152a8138402 req-9cd03dbe-b661-4ec9-811c-31de904d878f service nova] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Received event network-changed-53a4ce4d-f5b1-46ed-b4fe-db9454cef578 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.729084] env[63293]: DEBUG nova.compute.manager [req-092d11d0-5caa-4e36-840b-4152a8138402 req-9cd03dbe-b661-4ec9-811c-31de904d878f service nova] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Refreshing instance network info cache due to event network-changed-53a4ce4d-f5b1-46ed-b4fe-db9454cef578. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 857.729084] env[63293]: DEBUG oslo_concurrency.lockutils [req-092d11d0-5caa-4e36-840b-4152a8138402 req-9cd03dbe-b661-4ec9-811c-31de904d878f service nova] Acquiring lock "refresh_cache-9973dd7f-ca75-4db0-8d2b-f21b0311abeb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.729084] env[63293]: DEBUG oslo_concurrency.lockutils [req-092d11d0-5caa-4e36-840b-4152a8138402 req-9cd03dbe-b661-4ec9-811c-31de904d878f service nova] Acquired lock "refresh_cache-9973dd7f-ca75-4db0-8d2b-f21b0311abeb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.729084] env[63293]: DEBUG nova.network.neutron [req-092d11d0-5caa-4e36-840b-4152a8138402 req-9cd03dbe-b661-4ec9-811c-31de904d878f service nova] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Refreshing network info cache for port 53a4ce4d-f5b1-46ed-b4fe-db9454cef578 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 857.744838] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327704, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.929698] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327705, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.151716] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f3d281-d36c-e1f7-9200-e93da86b7f2d, 'name': SearchDatastore_Task, 'duration_secs': 0.059042} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.152067] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.152285] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.152523] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.152671] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.152852] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.153131] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20dce8ba-1c55-4dba-9eb1-e0f2f0b18965 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.161328] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.161504] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.162241] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9cfe55d-024a-4f43-afa8-831b24adf4f7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.167150] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 858.167150] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529a7d4d-003b-37ac-eda3-520dee458abb" [ 858.167150] env[63293]: _type = "Task" [ 858.167150] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.174107] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529a7d4d-003b-37ac-eda3-520dee458abb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.195578] env[63293]: DEBUG nova.scheduler.client.report [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.227909] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.245031] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327704, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.60223} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.245031] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 8ea7f3c3-8d73-4fb2-a848-4be4649e5660/8ea7f3c3-8d73-4fb2-a848-4be4649e5660.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 858.245031] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 858.245031] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30213266-ac94-4bda-b5ae-7cf6ba35f3ac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.251258] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 858.251258] env[63293]: value = "task-1327707" [ 858.251258] env[63293]: _type = "Task" [ 858.251258] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.259258] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327707, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.430022] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327705, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.445190] env[63293]: DEBUG nova.network.neutron [req-092d11d0-5caa-4e36-840b-4152a8138402 req-9cd03dbe-b661-4ec9-811c-31de904d878f service nova] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Updated VIF entry in instance network info cache for port 53a4ce4d-f5b1-46ed-b4fe-db9454cef578. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 858.445356] env[63293]: DEBUG nova.network.neutron [req-092d11d0-5caa-4e36-840b-4152a8138402 req-9cd03dbe-b661-4ec9-811c-31de904d878f service nova] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Updating instance_info_cache with network_info: [{"id": "53a4ce4d-f5b1-46ed-b4fe-db9454cef578", "address": "fa:16:3e:8b:8f:9f", "network": {"id": "b69cd9c0-f181-4f90-ac61-a0ebe2ea70b5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1584936849-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49f167005da748fcaf2497a5040a838c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53a4ce4d-f5", "ovs_interfaceid": "53a4ce4d-f5b1-46ed-b4fe-db9454cef578", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.678063] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529a7d4d-003b-37ac-eda3-520dee458abb, 'name': SearchDatastore_Task, 'duration_secs': 0.045841} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.678863] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54164fa5-c252-4360-8b34-a10b170e8c4c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.683902] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 858.683902] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]526e27c0-2933-7e7c-fb93-7d032a022f6f" [ 858.683902] env[63293]: _type = "Task" [ 858.683902] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.691160] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]526e27c0-2933-7e7c-fb93-7d032a022f6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.701013] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.701560] env[63293]: DEBUG nova.compute.manager [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 858.704152] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.089s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.704384] env[63293]: DEBUG nova.objects.instance [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lazy-loading 'resources' on Instance uuid 33d71260-26f1-482c-b93b-5f8e27c545f3 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.727145] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327706, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.760787] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327707, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067834} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.762028] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 858.762305] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832599b1-f445-451d-9ba4-563ee4403f9b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.784624] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 8ea7f3c3-8d73-4fb2-a848-4be4649e5660/8ea7f3c3-8d73-4fb2-a848-4be4649e5660.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 858.784944] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7315748-8c78-46b8-bf70-c1787c93d442 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.805354] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 858.805354] env[63293]: value = "task-1327709" [ 858.805354] env[63293]: _type = "Task" [ 858.805354] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.816239] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327709, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.929503] env[63293]: DEBUG oslo_vmware.api [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327705, 'name': PowerOnVM_Task, 'duration_secs': 1.192849} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.929788] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 858.929997] env[63293]: INFO nova.compute.manager [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Took 8.61 seconds to spawn the instance on the hypervisor. [ 858.930224] env[63293]: DEBUG nova.compute.manager [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.931008] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0995e225-d403-4b20-a32d-76b1e33a5ccf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.949436] env[63293]: DEBUG oslo_concurrency.lockutils [req-092d11d0-5caa-4e36-840b-4152a8138402 req-9cd03dbe-b661-4ec9-811c-31de904d878f service nova] Releasing lock "refresh_cache-9973dd7f-ca75-4db0-8d2b-f21b0311abeb" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.168299] env[63293]: DEBUG nova.network.neutron [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updating instance_info_cache with network_info: [{"id": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "address": "fa:16:3e:80:11:9c", "network": {"id": "9d5a3e58-567a-4c3f-9ad9-9fcbf81fced9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-949261114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1a30d37c1b47fd8670db0a9ef7020d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f0bfb1c-61", "ovs_interfaceid": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.195251] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]526e27c0-2933-7e7c-fb93-7d032a022f6f, 'name': SearchDatastore_Task, 'duration_secs': 0.008807} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.195529] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.195779] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 9973dd7f-ca75-4db0-8d2b-f21b0311abeb/9973dd7f-ca75-4db0-8d2b-f21b0311abeb.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 859.196053] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fe0071d7-4c7b-49d1-b597-2524fea2673b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.202599] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 859.202599] env[63293]: value = "task-1327710" [ 859.202599] env[63293]: _type = "Task" [ 859.202599] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.206811] env[63293]: DEBUG nova.compute.utils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.210588] env[63293]: DEBUG nova.compute.manager [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 859.210764] env[63293]: DEBUG nova.network.neutron [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 859.215846] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327710, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.227058] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327706, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.255614] env[63293]: DEBUG nova.policy [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7a950baa92f4d9ea77e5d6461582e0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec53642df3804e7190615487a426d4a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 859.321748] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.452743] env[63293]: INFO nova.compute.manager [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Took 28.30 seconds to build instance. [ 859.535263] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02019c3c-2a81-4fe8-8e50-b89dfbbe73a6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.543300] env[63293]: DEBUG nova.network.neutron [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Successfully created port: d41cf464-2396-4e22-80d1-3c1602939528 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.546925] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d33046-eb06-4fb7-ba5a-d909be660be3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.578730] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8886a8f-cca6-44bc-afae-1acfb6784913 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.586875] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa91a303-29ca-465b-b410-5379711f8b44 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.607060] env[63293]: DEBUG nova.compute.provider_tree [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.671523] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Releasing lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.672538] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updated the network info_cache for instance {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 859.672538] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.672538] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.672538] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.673461] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.673461] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.673461] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.673461] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63293) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 859.673461] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.714959] env[63293]: DEBUG nova.compute.manager [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 859.718310] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327710, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.728480] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327706, 'name': ReconfigVM_Task, 'duration_secs': 1.753253} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.728728] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Reconfigured VM instance instance-00000046 to attach disk [datastore1] f427933b-dd2c-49a9-b401-a647635a1eab/f427933b-dd2c-49a9-b401-a647635a1eab.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.730057] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1335e7d3-5fbb-4972-a02f-e2e811b54d72 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.740561] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for the task: (returnval){ [ 859.740561] env[63293]: value = "task-1327711" [ 859.740561] env[63293]: _type = "Task" [ 859.740561] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.753308] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327711, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.816899] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.957912] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9e5e5f05-ec21-407e-9115-c3b5eda15018 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.145s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.109862] env[63293]: DEBUG nova.scheduler.client.report [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.177620] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.198785] env[63293]: DEBUG oslo_concurrency.lockutils [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.199016] env[63293]: DEBUG oslo_concurrency.lockutils [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.199240] env[63293]: INFO nova.compute.manager [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Shelving [ 860.213181] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327710, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.641941} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.213427] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 9973dd7f-ca75-4db0-8d2b-f21b0311abeb/9973dd7f-ca75-4db0-8d2b-f21b0311abeb.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.213634] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.213893] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-85f82860-1335-47c1-b12d-1231817b74a3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.223367] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 860.223367] env[63293]: value = "task-1327712" [ 860.223367] env[63293]: _type = "Task" [ 860.223367] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.231979] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327712, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.249735] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327711, 'name': Rename_Task, 'duration_secs': 0.212013} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.249993] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.250292] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b173e91-92b8-482d-b627-c8595a28c216 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.256105] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for the task: (returnval){ [ 860.256105] env[63293]: value = "task-1327713" [ 860.256105] env[63293]: _type = "Task" [ 860.256105] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.264429] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327713, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.322067] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327709, 'name': ReconfigVM_Task, 'duration_secs': 1.160114} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.322067] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 8ea7f3c3-8d73-4fb2-a848-4be4649e5660/8ea7f3c3-8d73-4fb2-a848-4be4649e5660.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.322754] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd8e2b64-2bbd-4080-8a86-eff63b54d173 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.329771] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 860.329771] env[63293]: value = "task-1327714" [ 860.329771] env[63293]: _type = "Task" [ 860.329771] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.339985] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327714, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.599616] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Volume attach. Driver type: vmdk {{(pid=63293) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 860.599838] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283762', 'volume_id': '981997fd-c500-4d22-bd91-77b5fb478502', 'name': 'volume-981997fd-c500-4d22-bd91-77b5fb478502', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0977c79a-2ef3-4ec9-a0a0-de1ed5799723', 'attached_at': '', 'detached_at': '', 'volume_id': '981997fd-c500-4d22-bd91-77b5fb478502', 'serial': '981997fd-c500-4d22-bd91-77b5fb478502'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 860.600778] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8578eb-6497-4c53-bbca-c16ab02a66f2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.617333] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.913s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.619747] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.918s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.621406] env[63293]: INFO nova.compute.claims [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 860.624890] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6019ae3-cf93-4e9b-9f44-04a0c7fe9e98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.653327] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] volume-981997fd-c500-4d22-bd91-77b5fb478502/volume-981997fd-c500-4d22-bd91-77b5fb478502.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.654548] env[63293]: INFO nova.scheduler.client.report [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted allocations for instance 33d71260-26f1-482c-b93b-5f8e27c545f3 [ 860.655656] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc07de14-18c8-41ee-b0db-a4d3170b808c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.677373] env[63293]: DEBUG oslo_vmware.api [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 860.677373] env[63293]: value = "task-1327715" [ 860.677373] env[63293]: _type = "Task" [ 860.677373] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.685706] env[63293]: DEBUG oslo_vmware.api [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327715, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.707037] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 860.707178] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-229f6937-e768-4e56-b017-7d6e35c62517 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.713329] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 860.713329] env[63293]: value = "task-1327716" [ 860.713329] env[63293]: _type = "Task" [ 860.713329] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.722736] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327716, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.725080] env[63293]: DEBUG nova.compute.manager [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 860.735254] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327712, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057176} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.736025] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.736547] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0943c6fe-a346-47b4-b3cb-a24f61b43658 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.760149] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 9973dd7f-ca75-4db0-8d2b-f21b0311abeb/9973dd7f-ca75-4db0-8d2b-f21b0311abeb.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.762571] env[63293]: DEBUG nova.virt.hardware [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.762776] env[63293]: DEBUG nova.virt.hardware [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.762949] env[63293]: DEBUG nova.virt.hardware [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.763143] env[63293]: DEBUG nova.virt.hardware [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.763296] env[63293]: DEBUG nova.virt.hardware [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.763491] env[63293]: DEBUG nova.virt.hardware [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.763714] env[63293]: DEBUG nova.virt.hardware [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.763893] env[63293]: DEBUG nova.virt.hardware [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.764176] env[63293]: DEBUG nova.virt.hardware [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.764252] env[63293]: DEBUG nova.virt.hardware [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.764390] env[63293]: DEBUG nova.virt.hardware [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.765047] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-596a5669-fa6a-4ea2-b3c1-e4f740592fa7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.780470] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fa2b9a-37c9-4445-88df-288bbe0ca6b6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.795928] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36406359-568e-4b47-830d-30ddc431755b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.801820] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327713, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.801820] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 860.801820] env[63293]: value = "task-1327717" [ 860.801820] env[63293]: _type = "Task" [ 860.801820] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.819958] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327717, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.839799] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327714, 'name': Rename_Task, 'duration_secs': 0.161025} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.840106] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.840376] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f8061ae-51f7-4714-b4db-50ab3b7b74ff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.847569] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 860.847569] env[63293]: value = "task-1327718" [ 860.847569] env[63293]: _type = "Task" [ 860.847569] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.859920] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327718, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.177887] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53efe829-1519-4ca1-bf08-02f608986585 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "33d71260-26f1-482c-b93b-5f8e27c545f3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.677s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.192075] env[63293]: DEBUG oslo_vmware.api [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327715, 'name': ReconfigVM_Task, 'duration_secs': 0.475338} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.192075] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Reconfigured VM instance instance-00000039 to attach disk [datastore2] volume-981997fd-c500-4d22-bd91-77b5fb478502/volume-981997fd-c500-4d22-bd91-77b5fb478502.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.196792] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73f4f206-7c58-40e4-a1b8-862b18d4e7fe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.210612] env[63293]: DEBUG nova.compute.manager [req-be24d720-1447-4d80-b718-fce9608786a8 req-717f36e3-996b-48c9-bd2b-9e4c50b9bddc service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Received event network-vif-plugged-d41cf464-2396-4e22-80d1-3c1602939528 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.210821] env[63293]: DEBUG oslo_concurrency.lockutils [req-be24d720-1447-4d80-b718-fce9608786a8 req-717f36e3-996b-48c9-bd2b-9e4c50b9bddc service nova] Acquiring lock "d753598d-a92a-4515-9ad1-d386294f7a99-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.211119] env[63293]: DEBUG oslo_concurrency.lockutils [req-be24d720-1447-4d80-b718-fce9608786a8 req-717f36e3-996b-48c9-bd2b-9e4c50b9bddc service nova] Lock "d753598d-a92a-4515-9ad1-d386294f7a99-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.211315] env[63293]: DEBUG oslo_concurrency.lockutils [req-be24d720-1447-4d80-b718-fce9608786a8 req-717f36e3-996b-48c9-bd2b-9e4c50b9bddc service nova] Lock "d753598d-a92a-4515-9ad1-d386294f7a99-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.211552] env[63293]: DEBUG nova.compute.manager [req-be24d720-1447-4d80-b718-fce9608786a8 req-717f36e3-996b-48c9-bd2b-9e4c50b9bddc service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] No waiting events found dispatching network-vif-plugged-d41cf464-2396-4e22-80d1-3c1602939528 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 861.211776] env[63293]: WARNING nova.compute.manager [req-be24d720-1447-4d80-b718-fce9608786a8 req-717f36e3-996b-48c9-bd2b-9e4c50b9bddc service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Received unexpected event network-vif-plugged-d41cf464-2396-4e22-80d1-3c1602939528 for instance with vm_state building and task_state spawning. [ 861.217956] env[63293]: DEBUG oslo_vmware.api [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 861.217956] env[63293]: value = "task-1327719" [ 861.217956] env[63293]: _type = "Task" [ 861.217956] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.228948] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327716, 'name': PowerOffVM_Task, 'duration_secs': 0.185521} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.232088] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 861.232487] env[63293]: DEBUG oslo_vmware.api [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327719, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.233138] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c431af-b304-4879-921a-34806655b787 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.251726] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d33eae9c-4948-4562-974c-5baee7454e98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.275371] env[63293]: DEBUG oslo_vmware.api [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327713, 'name': PowerOnVM_Task, 'duration_secs': 0.604266} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.275371] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.275371] env[63293]: INFO nova.compute.manager [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Took 13.55 seconds to spawn the instance on the hypervisor. [ 861.275371] env[63293]: DEBUG nova.compute.manager [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.276119] env[63293]: DEBUG nova.network.neutron [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Successfully updated port: d41cf464-2396-4e22-80d1-3c1602939528 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.278365] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93f0b89-d91a-4e2d-876d-9ace3ac1c951 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.314469] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327717, 'name': ReconfigVM_Task, 'duration_secs': 0.383671} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.315467] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 9973dd7f-ca75-4db0-8d2b-f21b0311abeb/9973dd7f-ca75-4db0-8d2b-f21b0311abeb.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.316896] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d332a03e-f4e4-4237-858f-53a94fdad5f7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.331636] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 861.331636] env[63293]: value = "task-1327720" [ 861.331636] env[63293]: _type = "Task" [ 861.331636] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.341375] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327720, 'name': Rename_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.357788] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327718, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.732419] env[63293]: DEBUG oslo_vmware.api [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327719, 'name': ReconfigVM_Task, 'duration_secs': 0.149968} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.732749] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283762', 'volume_id': '981997fd-c500-4d22-bd91-77b5fb478502', 'name': 'volume-981997fd-c500-4d22-bd91-77b5fb478502', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0977c79a-2ef3-4ec9-a0a0-de1ed5799723', 'attached_at': '', 'detached_at': '', 'volume_id': '981997fd-c500-4d22-bd91-77b5fb478502', 'serial': '981997fd-c500-4d22-bd91-77b5fb478502'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 861.763502] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Creating Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 861.764420] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9d1fa837-9451-4569-8777-2323bec6785a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.771796] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 861.771796] env[63293]: value = "task-1327721" [ 861.771796] env[63293]: _type = "Task" [ 861.771796] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.782304] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "refresh_cache-d753598d-a92a-4515-9ad1-d386294f7a99" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.782495] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "refresh_cache-d753598d-a92a-4515-9ad1-d386294f7a99" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.782553] env[63293]: DEBUG nova.network.neutron [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.783646] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327721, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.796499] env[63293]: INFO nova.compute.manager [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Took 34.02 seconds to build instance. [ 861.842894] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327720, 'name': Rename_Task, 'duration_secs': 0.144788} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.843113] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.845712] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cebf7dca-2f9e-413b-8d31-902e1ed74767 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.853150] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 861.853150] env[63293]: value = "task-1327722" [ 861.853150] env[63293]: _type = "Task" [ 861.853150] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.859422] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327718, 'name': PowerOnVM_Task, 'duration_secs': 0.979209} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.862243] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.862474] env[63293]: INFO nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Took 9.03 seconds to spawn the instance on the hypervisor. [ 861.862652] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.863762] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34fcda62-f5f4-4585-85a7-fc562065f0a1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.869011] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327722, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.926359] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff84420b-8365-406c-b92e-c7a3ab30cb00 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.935052] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfbf88fb-5dd1-4957-98d1-b12ddee2dae7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.968526] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63162879-1503-4c43-a67a-e7db8f86ed30 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.976532] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907efc9e-ea17-4ece-9fea-fd675bf48be6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.990147] env[63293]: DEBUG nova.compute.provider_tree [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.190808] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1b6bfde3-95e5-4c3a-b3c4-6de98d2109a5 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.191164] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1b6bfde3-95e5-4c3a-b3c4-6de98d2109a5 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.191603] env[63293]: DEBUG nova.objects.instance [None req-1b6bfde3-95e5-4c3a-b3c4-6de98d2109a5 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'flavor' on Instance uuid 65c2f72d-6bbe-45ac-8efc-401e0664390f {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.281157] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327721, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.298966] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff1298f7-1f2c-4eac-8c4d-2934d22e4e08 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "f427933b-dd2c-49a9-b401-a647635a1eab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.447s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.316420] env[63293]: DEBUG nova.network.neutron [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 862.364958] env[63293]: DEBUG oslo_vmware.api [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327722, 'name': PowerOnVM_Task, 'duration_secs': 0.467744} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.364958] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.365122] env[63293]: INFO nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Took 7.07 seconds to spawn the instance on the hypervisor. [ 862.365280] env[63293]: DEBUG nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.366066] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb503675-ccd2-4315-a2e2-35dfdeddab35 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.387095] env[63293]: INFO nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Took 28.19 seconds to build instance. [ 862.463735] env[63293]: DEBUG nova.network.neutron [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Updating instance_info_cache with network_info: [{"id": "d41cf464-2396-4e22-80d1-3c1602939528", "address": "fa:16:3e:1e:ad:a7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd41cf464-23", "ovs_interfaceid": "d41cf464-2396-4e22-80d1-3c1602939528", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.494592] env[63293]: DEBUG nova.scheduler.client.report [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.695519] env[63293]: DEBUG nova.objects.instance [None req-1b6bfde3-95e5-4c3a-b3c4-6de98d2109a5 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'pci_requests' on Instance uuid 65c2f72d-6bbe-45ac-8efc-401e0664390f {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.770199] env[63293]: DEBUG nova.objects.instance [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'flavor' on Instance uuid 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.782335] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327721, 'name': CreateSnapshot_Task, 'duration_secs': 0.578812} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.782625] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Created Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 862.784022] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf4c2e2-fa34-4246-9943-d9cf03cd2e30 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.884228] env[63293]: INFO nova.compute.manager [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Took 25.81 seconds to build instance. [ 862.889416] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.787s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.956667] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "f427933b-dd2c-49a9-b401-a647635a1eab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.956952] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "f427933b-dd2c-49a9-b401-a647635a1eab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.957196] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "f427933b-dd2c-49a9-b401-a647635a1eab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.957379] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "f427933b-dd2c-49a9-b401-a647635a1eab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.957549] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "f427933b-dd2c-49a9-b401-a647635a1eab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.959778] env[63293]: INFO nova.compute.manager [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Terminating instance [ 862.961650] env[63293]: DEBUG nova.compute.manager [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 862.961862] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 862.962716] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933f3ad1-854b-4771-bcab-bd5aa798d241 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.967781] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "refresh_cache-d753598d-a92a-4515-9ad1-d386294f7a99" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.968057] env[63293]: DEBUG nova.compute.manager [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Instance network_info: |[{"id": "d41cf464-2396-4e22-80d1-3c1602939528", "address": "fa:16:3e:1e:ad:a7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd41cf464-23", "ovs_interfaceid": "d41cf464-2396-4e22-80d1-3c1602939528", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 862.968499] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:ad:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91c1da19-ab68-4127-bacd-accbaff19651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd41cf464-2396-4e22-80d1-3c1602939528', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.975837] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Creating folder: Project (ec53642df3804e7190615487a426d4a6). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.978458] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28d89fe4-2a7d-4b3a-ada6-54ec819dbba6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.980139] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 862.980373] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06514dc5-e6a2-4ab1-8f85-7e76e4862f9a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.988504] env[63293]: DEBUG oslo_vmware.api [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for the task: (returnval){ [ 862.988504] env[63293]: value = "task-1327723" [ 862.988504] env[63293]: _type = "Task" [ 862.988504] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.992821] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Created folder: Project (ec53642df3804e7190615487a426d4a6) in parent group-v283678. [ 862.993019] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Creating folder: Instances. Parent ref: group-v283765. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.993570] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c3aee704-2eda-405d-a1d8-ff7867b484f0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.998586] env[63293]: DEBUG oslo_vmware.api [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327723, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.000288] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.000804] env[63293]: DEBUG nova.compute.manager [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 863.003446] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.980s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.003667] env[63293]: DEBUG nova.objects.instance [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Lazy-loading 'resources' on Instance uuid 1e01a75d-122d-4122-9fed-4164d64d4ee8 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.007442] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Created folder: Instances in parent group-v283765. [ 863.007851] env[63293]: DEBUG oslo.service.loopingcall [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.007913] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 863.008083] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a20db765-1aef-4c38-b49b-ddb558343520 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.030334] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.030334] env[63293]: value = "task-1327726" [ 863.030334] env[63293]: _type = "Task" [ 863.030334] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.038526] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327726, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.198273] env[63293]: DEBUG nova.objects.base [None req-1b6bfde3-95e5-4c3a-b3c4-6de98d2109a5 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Object Instance<65c2f72d-6bbe-45ac-8efc-401e0664390f> lazy-loaded attributes: flavor,pci_requests {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 863.198729] env[63293]: DEBUG nova.network.neutron [None req-1b6bfde3-95e5-4c3a-b3c4-6de98d2109a5 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 863.276066] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec521080-e676-46d6-af8c-25fea19b99f9 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.283s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.285597] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1b6bfde3-95e5-4c3a-b3c4-6de98d2109a5 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.094s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.302793] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Creating linked-clone VM from snapshot {{(pid=63293) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 863.303738] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a6ffdc7b-3789-4678-a8c0-f11468f4780d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.313584] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 863.313584] env[63293]: value = "task-1327727" [ 863.313584] env[63293]: _type = "Task" [ 863.313584] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.321973] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327727, 'name': CloneVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.334511] env[63293]: DEBUG nova.compute.manager [req-0009e7bc-1e83-4b72-b5de-e3a1baa80d4e req-2b9b40e7-33b1-4d48-8499-41c639b0988a service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Received event network-changed-d41cf464-2396-4e22-80d1-3c1602939528 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.334713] env[63293]: DEBUG nova.compute.manager [req-0009e7bc-1e83-4b72-b5de-e3a1baa80d4e req-2b9b40e7-33b1-4d48-8499-41c639b0988a service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Refreshing instance network info cache due to event network-changed-d41cf464-2396-4e22-80d1-3c1602939528. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 863.335181] env[63293]: DEBUG oslo_concurrency.lockutils [req-0009e7bc-1e83-4b72-b5de-e3a1baa80d4e req-2b9b40e7-33b1-4d48-8499-41c639b0988a service nova] Acquiring lock "refresh_cache-d753598d-a92a-4515-9ad1-d386294f7a99" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.335181] env[63293]: DEBUG oslo_concurrency.lockutils [req-0009e7bc-1e83-4b72-b5de-e3a1baa80d4e req-2b9b40e7-33b1-4d48-8499-41c639b0988a service nova] Acquired lock "refresh_cache-d753598d-a92a-4515-9ad1-d386294f7a99" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.335329] env[63293]: DEBUG nova.network.neutron [req-0009e7bc-1e83-4b72-b5de-e3a1baa80d4e req-2b9b40e7-33b1-4d48-8499-41c639b0988a service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Refreshing network info cache for port d41cf464-2396-4e22-80d1-3c1602939528 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 863.387403] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fc064a7-74f2-4077-8989-852d16d062d7 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.258s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.490443] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.490719] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.490917] env[63293]: DEBUG nova.compute.manager [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.491924] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0cd75d9-e69d-4c4b-903c-5d85ca4edf52 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.501963] env[63293]: DEBUG nova.compute.manager [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63293) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 863.502567] env[63293]: DEBUG nova.objects.instance [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'flavor' on Instance uuid 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.511038] env[63293]: DEBUG nova.compute.utils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 863.512041] env[63293]: DEBUG oslo_vmware.api [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327723, 'name': PowerOffVM_Task, 'duration_secs': 0.268788} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.512744] env[63293]: DEBUG nova.compute.manager [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Not allocating networking since 'none' was specified. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 863.513260] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.513260] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 863.513484] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a361e287-53a1-4065-8199-bf5bedf5d546 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.544151] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327726, 'name': CreateVM_Task, 'duration_secs': 0.378535} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.546641] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 863.547850] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.548042] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.548466] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 863.548961] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2611080-8468-46ba-8026-bdc14dd18976 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.553958] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 863.553958] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524b8591-bc2b-33c0-5e0e-028cb4ac8ebe" [ 863.553958] env[63293]: _type = "Task" [ 863.553958] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.565379] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524b8591-bc2b-33c0-5e0e-028cb4ac8ebe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.682468] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 863.682468] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 863.682649] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Deleting the datastore file [datastore1] f427933b-dd2c-49a9-b401-a647635a1eab {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.682910] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c40eed74-c720-4fbd-9160-f7015d0ff4f7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.689476] env[63293]: DEBUG oslo_vmware.api [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for the task: (returnval){ [ 863.689476] env[63293]: value = "task-1327729" [ 863.689476] env[63293]: _type = "Task" [ 863.689476] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.697667] env[63293]: DEBUG oslo_vmware.api [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.784628] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f47e9e9-d224-4614-8fbd-77c00f4433d2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.795251] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358c0fe2-a36c-49cd-9df6-9293c3ac4fc6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.832023] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0847715-169f-4ec9-95e0-c926c964743e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.841081] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327727, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.848233] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43cd7f67-4e0f-4288-9626-292bf88851d5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.859586] env[63293]: DEBUG nova.compute.provider_tree [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.015448] env[63293]: DEBUG nova.compute.manager [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 864.022039] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.022039] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-27f72f78-4617-4e8b-83e8-06e5ce67329e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.028330] env[63293]: DEBUG oslo_vmware.api [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 864.028330] env[63293]: value = "task-1327730" [ 864.028330] env[63293]: _type = "Task" [ 864.028330] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.036354] env[63293]: DEBUG oslo_vmware.api [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327730, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.066471] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524b8591-bc2b-33c0-5e0e-028cb4ac8ebe, 'name': SearchDatastore_Task, 'duration_secs': 0.012523} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.067629] env[63293]: DEBUG nova.network.neutron [req-0009e7bc-1e83-4b72-b5de-e3a1baa80d4e req-2b9b40e7-33b1-4d48-8499-41c639b0988a service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Updated VIF entry in instance network info cache for port d41cf464-2396-4e22-80d1-3c1602939528. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 864.068180] env[63293]: DEBUG nova.network.neutron [req-0009e7bc-1e83-4b72-b5de-e3a1baa80d4e req-2b9b40e7-33b1-4d48-8499-41c639b0988a service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Updating instance_info_cache with network_info: [{"id": "d41cf464-2396-4e22-80d1-3c1602939528", "address": "fa:16:3e:1e:ad:a7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd41cf464-23", "ovs_interfaceid": "d41cf464-2396-4e22-80d1-3c1602939528", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.069662] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.070075] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.070576] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.070965] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.071347] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.072246] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40abde16-c551-4efd-a7cd-77d91a3a1a25 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.085127] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.085499] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 864.086571] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1079eb6-45a7-4c07-95ec-d9f04c544f0a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.091800] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 864.091800] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528527a5-db82-674b-4fe2-b63191d1e7bb" [ 864.091800] env[63293]: _type = "Task" [ 864.091800] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.099854] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528527a5-db82-674b-4fe2-b63191d1e7bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.199542] env[63293]: DEBUG oslo_vmware.api [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Task: {'id': task-1327729, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216893} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.199854] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.200019] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 864.200213] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 864.200438] env[63293]: INFO nova.compute.manager [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Took 1.24 seconds to destroy the instance on the hypervisor. [ 864.200698] env[63293]: DEBUG oslo.service.loopingcall [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.200897] env[63293]: DEBUG nova.compute.manager [-] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 864.200996] env[63293]: DEBUG nova.network.neutron [-] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 864.227736] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.227985] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.228223] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.228469] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.228626] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.230994] env[63293]: INFO nova.compute.manager [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Terminating instance [ 864.232781] env[63293]: DEBUG nova.compute.manager [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 864.232971] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.233799] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04013141-058f-4ee2-9b12-90ac363b5ef8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.241263] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.241499] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0a4424f-23d3-4e89-8640-71c8f72ba682 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.247275] env[63293]: DEBUG oslo_vmware.api [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 864.247275] env[63293]: value = "task-1327731" [ 864.247275] env[63293]: _type = "Task" [ 864.247275] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.255237] env[63293]: DEBUG oslo_vmware.api [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.339557] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327727, 'name': CloneVM_Task} progress is 95%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.362848] env[63293]: DEBUG nova.scheduler.client.report [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.539461] env[63293]: DEBUG oslo_vmware.api [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327730, 'name': PowerOffVM_Task, 'duration_secs': 0.413141} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.539741] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.539915] env[63293]: DEBUG nova.compute.manager [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 864.540710] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba3cb73-3ee4-4723-ba02-37526149a1e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.545383] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.547017] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.547017] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.547017] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.547017] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.551458] env[63293]: INFO nova.compute.manager [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Terminating instance [ 864.553414] env[63293]: DEBUG nova.compute.manager [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 864.553604] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.554361] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd329959-0c05-4c68-842f-8f1a8ccc4638 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.561631] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.561631] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cba39e18-263f-44a1-8c84-1d59f68fb56a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.572330] env[63293]: DEBUG oslo_vmware.api [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 864.572330] env[63293]: value = "task-1327732" [ 864.572330] env[63293]: _type = "Task" [ 864.572330] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.573121] env[63293]: DEBUG oslo_concurrency.lockutils [req-0009e7bc-1e83-4b72-b5de-e3a1baa80d4e req-2b9b40e7-33b1-4d48-8499-41c639b0988a service nova] Releasing lock "refresh_cache-d753598d-a92a-4515-9ad1-d386294f7a99" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.578432] env[63293]: DEBUG oslo_vmware.api [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327732, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.605767] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528527a5-db82-674b-4fe2-b63191d1e7bb, 'name': SearchDatastore_Task, 'duration_secs': 0.009653} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.606664] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3a13de6-1407-4f61-9ff4-dbc96b2fe5c3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.613159] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 864.613159] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520bb8ed-1c41-f95a-31ec-691da60ec115" [ 864.613159] env[63293]: _type = "Task" [ 864.613159] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.625974] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520bb8ed-1c41-f95a-31ec-691da60ec115, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.758191] env[63293]: DEBUG oslo_vmware.api [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327731, 'name': PowerOffVM_Task, 'duration_secs': 0.192266} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.758550] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.758733] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 864.759041] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9a9a2a1-b70c-406e-b8cf-481233483e94 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.813660] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.813955] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.814166] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Deleting the datastore file [datastore2] 8ea7f3c3-8d73-4fb2-a848-4be4649e5660 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.814432] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c92d85ab-d543-4462-93d0-02fc669a357d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.821244] env[63293]: DEBUG oslo_vmware.api [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 864.821244] env[63293]: value = "task-1327734" [ 864.821244] env[63293]: _type = "Task" [ 864.821244] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.830288] env[63293]: DEBUG oslo_vmware.api [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327734, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.839143] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327727, 'name': CloneVM_Task, 'duration_secs': 1.390912} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.839446] env[63293]: INFO nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Created linked-clone VM from snapshot [ 864.840214] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e294eafd-0661-47a6-8446-cc341b78744b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.847345] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Uploading image 46aa5632-364b-4607-b217-b473d1ef29ad {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 864.869061] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.870986] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.575s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.872471] env[63293]: INFO nova.compute.claims [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.876955] env[63293]: DEBUG oslo_vmware.rw_handles [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 864.876955] env[63293]: value = "vm-283768" [ 864.876955] env[63293]: _type = "VirtualMachine" [ 864.876955] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 864.876955] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c6ed36dd-9e1f-40b2-950f-45b5db619597 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.884309] env[63293]: DEBUG oslo_vmware.rw_handles [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lease: (returnval){ [ 864.884309] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a3a90a-3f36-76be-2402-2e426b3de7c4" [ 864.884309] env[63293]: _type = "HttpNfcLease" [ 864.884309] env[63293]: } obtained for exporting VM: (result){ [ 864.884309] env[63293]: value = "vm-283768" [ 864.884309] env[63293]: _type = "VirtualMachine" [ 864.884309] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 864.884309] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the lease: (returnval){ [ 864.884309] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a3a90a-3f36-76be-2402-2e426b3de7c4" [ 864.884309] env[63293]: _type = "HttpNfcLease" [ 864.884309] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 864.890502] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 864.890502] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a3a90a-3f36-76be-2402-2e426b3de7c4" [ 864.890502] env[63293]: _type = "HttpNfcLease" [ 864.890502] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 864.897800] env[63293]: INFO nova.scheduler.client.report [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Deleted allocations for instance 1e01a75d-122d-4122-9fed-4164d64d4ee8 [ 865.032032] env[63293]: DEBUG nova.compute.manager [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 865.054484] env[63293]: DEBUG nova.virt.hardware [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 865.056315] env[63293]: DEBUG nova.virt.hardware [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 865.056315] env[63293]: DEBUG nova.virt.hardware [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.056315] env[63293]: DEBUG nova.virt.hardware [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 865.056315] env[63293]: DEBUG nova.virt.hardware [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.056315] env[63293]: DEBUG nova.virt.hardware [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 865.056315] env[63293]: DEBUG nova.virt.hardware [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 865.056315] env[63293]: DEBUG nova.virt.hardware [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 865.056315] env[63293]: DEBUG nova.virt.hardware [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 865.056912] env[63293]: DEBUG nova.virt.hardware [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 865.056912] env[63293]: DEBUG nova.virt.hardware [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 865.059664] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a6968f-d633-4445-9347-0d0542042ded {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.062878] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4a1be7a-09c1-49a3-8eb0-220b601f10f1 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.572s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.070696] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45e3405-4558-422b-a7df-da7795527ec7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.088561] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.094268] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Creating folder: Project (48792eaf58f74e308bb6512690e4c90f). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.097754] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f459248-76a9-4d59-a210-e6105519f0fc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.099422] env[63293]: DEBUG oslo_vmware.api [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327732, 'name': PowerOffVM_Task, 'duration_secs': 0.219826} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.099676] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 865.099847] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 865.100485] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b8b72cdd-de42-465c-b1ee-5b9137d30a3d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.108934] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Created folder: Project (48792eaf58f74e308bb6512690e4c90f) in parent group-v283678. [ 865.109137] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Creating folder: Instances. Parent ref: group-v283769. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.109397] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-088b1adf-2f32-4801-8ec4-a3ca71b34e87 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.117703] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Created folder: Instances in parent group-v283769. [ 865.117958] env[63293]: DEBUG oslo.service.loopingcall [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.120934] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 865.121204] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520bb8ed-1c41-f95a-31ec-691da60ec115, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.121405] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4fecb262-d60b-4a77-8725-18c14e764f15 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.137636] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.137636] env[63293]: value = "task-1327739" [ 865.137636] env[63293]: _type = "Task" [ 865.137636] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.145266] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327739, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.166195] env[63293]: DEBUG nova.network.neutron [-] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.182751] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 865.182994] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 865.183183] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Deleting the datastore file [datastore2] 9973dd7f-ca75-4db0-8d2b-f21b0311abeb {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.184659] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f6a52f9-8f54-4a8d-92fe-085fc59f21ca {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.191032] env[63293]: DEBUG oslo_vmware.api [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for the task: (returnval){ [ 865.191032] env[63293]: value = "task-1327740" [ 865.191032] env[63293]: _type = "Task" [ 865.191032] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.199445] env[63293]: DEBUG oslo_vmware.api [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.286948] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.287216] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.287417] env[63293]: DEBUG nova.objects.instance [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'flavor' on Instance uuid 65c2f72d-6bbe-45ac-8efc-401e0664390f {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.331437] env[63293]: DEBUG oslo_vmware.api [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327734, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145104} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.331712] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.331899] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 865.332087] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.332261] env[63293]: INFO nova.compute.manager [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Took 1.10 seconds to destroy the instance on the hypervisor. [ 865.332502] env[63293]: DEBUG oslo.service.loopingcall [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.332696] env[63293]: DEBUG nova.compute.manager [-] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 865.332794] env[63293]: DEBUG nova.network.neutron [-] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 865.395958] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 865.395958] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a3a90a-3f36-76be-2402-2e426b3de7c4" [ 865.395958] env[63293]: _type = "HttpNfcLease" [ 865.395958] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 865.396692] env[63293]: DEBUG oslo_vmware.rw_handles [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 865.396692] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a3a90a-3f36-76be-2402-2e426b3de7c4" [ 865.396692] env[63293]: _type = "HttpNfcLease" [ 865.396692] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 865.397530] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acdfa70-dc28-49df-80b9-e422a2dee484 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.405806] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0b0a7ac-2b6f-4e64-bb2f-fc7b4b635862 tempest-ServerAddressesTestJSON-1889171944 tempest-ServerAddressesTestJSON-1889171944-project-member] Lock "1e01a75d-122d-4122-9fed-4164d64d4ee8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.362s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.407566] env[63293]: DEBUG nova.compute.manager [req-5b7fb93f-8ee6-40e2-bbad-62e2475edc1d req-e03edb75-5a04-4bb4-989e-3987e2852bb1 service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Received event network-vif-deleted-31fd0cd3-702f-4cea-860d-fc21f02effb2 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.407914] env[63293]: DEBUG nova.compute.manager [req-5b7fb93f-8ee6-40e2-bbad-62e2475edc1d req-e03edb75-5a04-4bb4-989e-3987e2852bb1 service nova] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Received event network-vif-deleted-b6279a11-248c-4e66-9df7-aa175821b8f1 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.412673] env[63293]: DEBUG oslo_vmware.rw_handles [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d134c-1e3f-e738-6d20-c536c1bd38a7/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 865.412887] env[63293]: DEBUG oslo_vmware.rw_handles [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d134c-1e3f-e738-6d20-c536c1bd38a7/disk-0.vmdk for reading. {{(pid=63293) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 865.523600] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c24fc9ae-8b3a-4f3c-bdf4-ce089c4c851f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.624604] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520bb8ed-1c41-f95a-31ec-691da60ec115, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.647694] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327739, 'name': CreateVM_Task, 'duration_secs': 0.428725} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.651021] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.651021] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.651021] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.651021] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 865.651021] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ff39caa-1501-4ee3-b0ed-7ab31128b9a1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.653796] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 865.653796] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5240634a-84d6-f3ef-c02e-f47559971b0a" [ 865.653796] env[63293]: _type = "Task" [ 865.653796] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.661934] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5240634a-84d6-f3ef-c02e-f47559971b0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.668558] env[63293]: INFO nova.compute.manager [-] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Took 1.47 seconds to deallocate network for instance. [ 865.700605] env[63293]: DEBUG oslo_vmware.api [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Task: {'id': task-1327740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161114} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.701054] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.701352] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 865.701643] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.701960] env[63293]: INFO nova.compute.manager [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Took 1.15 seconds to destroy the instance on the hypervisor. [ 865.702321] env[63293]: DEBUG oslo.service.loopingcall [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.702627] env[63293]: DEBUG nova.compute.manager [-] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 865.703805] env[63293]: DEBUG nova.network.neutron [-] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 865.890222] env[63293]: DEBUG nova.objects.instance [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'pci_requests' on Instance uuid 65c2f72d-6bbe-45ac-8efc-401e0664390f {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.082410] env[63293]: DEBUG nova.network.neutron [-] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.127347] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520bb8ed-1c41-f95a-31ec-691da60ec115, 'name': SearchDatastore_Task, 'duration_secs': 1.389337} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.130207] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.130502] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] d753598d-a92a-4515-9ad1-d386294f7a99/d753598d-a92a-4515-9ad1-d386294f7a99.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.131228] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0221e21c-8cf4-4b1b-a40c-617ce7ac36d9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.138444] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 866.138444] env[63293]: value = "task-1327741" [ 866.138444] env[63293]: _type = "Task" [ 866.138444] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.148893] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327741, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.165949] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5240634a-84d6-f3ef-c02e-f47559971b0a, 'name': SearchDatastore_Task, 'duration_secs': 0.348272} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.166627] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.166887] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 866.166963] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.167171] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.167377] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.170614] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e2b314f5-9f36-40f4-945d-12b14523bc11 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.175326] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.178755] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.179049] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.179728] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a02605d-1714-40a5-9905-8cc20eb574f5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.185169] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 866.185169] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522069b5-129f-0150-e982-f8668a8b62f6" [ 866.185169] env[63293]: _type = "Task" [ 866.185169] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.196245] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522069b5-129f-0150-e982-f8668a8b62f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.204879] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a303a4-5199-45a1-8bc9-426907b3f40d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.214754] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f9e7a0-4d8f-439f-836a-9db570660eff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.247589] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d287bb-7b3e-4bc4-adc9-01326e03598e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.255832] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a3c1c1f-4d21-4c37-8378-47692d705aed {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.269555] env[63293]: DEBUG nova.compute.provider_tree [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.393789] env[63293]: DEBUG nova.objects.base [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Object Instance<65c2f72d-6bbe-45ac-8efc-401e0664390f> lazy-loaded attributes: flavor,pci_requests {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 866.394200] env[63293]: DEBUG nova.network.neutron [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 866.405438] env[63293]: DEBUG nova.objects.instance [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'flavor' on Instance uuid 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.463032] env[63293]: DEBUG nova.network.neutron [-] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.478699] env[63293]: DEBUG nova.policy [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2312d48038d7474481a0709267645780', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '866b347100454019b07f63922b995bb6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 866.587279] env[63293]: INFO nova.compute.manager [-] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Took 1.25 seconds to deallocate network for instance. [ 866.650193] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327741, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484604} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.650803] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] d753598d-a92a-4515-9ad1-d386294f7a99/d753598d-a92a-4515-9ad1-d386294f7a99.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.651143] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.651667] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f04c9cd-bd19-4b1f-ae15-f5b961f81eaf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.658347] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 866.658347] env[63293]: value = "task-1327742" [ 866.658347] env[63293]: _type = "Task" [ 866.658347] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.669458] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327742, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.697056] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522069b5-129f-0150-e982-f8668a8b62f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009275} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.697920] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12021a9d-5d10-42ed-9a0b-0fde423d1686 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.703791] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 866.703791] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a97e83-f7ee-eefa-de32-e60388a4d278" [ 866.703791] env[63293]: _type = "Task" [ 866.703791] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.712054] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a97e83-f7ee-eefa-de32-e60388a4d278, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.773028] env[63293]: DEBUG nova.scheduler.client.report [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.915683] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.915683] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquired lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.915683] env[63293]: DEBUG nova.network.neutron [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 866.915683] env[63293]: DEBUG nova.objects.instance [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'info_cache' on Instance uuid 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.965284] env[63293]: INFO nova.compute.manager [-] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Took 1.26 seconds to deallocate network for instance. [ 867.016444] env[63293]: DEBUG nova.network.neutron [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Successfully created port: 50288bf0-633c-4e61-a850-e343398936c4 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.097029] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.168714] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327742, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086496} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.169045] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.169952] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e08472-c7b2-4d03-83d7-e1a8758c85b8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.198482] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] d753598d-a92a-4515-9ad1-d386294f7a99/d753598d-a92a-4515-9ad1-d386294f7a99.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.198881] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6cc4c640-0139-4508-aea8-0be10c112fc6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.225536] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a97e83-f7ee-eefa-de32-e60388a4d278, 'name': SearchDatastore_Task, 'duration_secs': 0.012223} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.226954] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.227318] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 828a1165-3829-431c-9edb-3a5d3a8054ec/828a1165-3829-431c-9edb-3a5d3a8054ec.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 867.227649] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 867.227649] env[63293]: value = "task-1327743" [ 867.227649] env[63293]: _type = "Task" [ 867.227649] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.227846] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-106f5188-de41-415b-99dc-84cbb5728487 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.238196] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327743, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.239576] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 867.239576] env[63293]: value = "task-1327744" [ 867.239576] env[63293]: _type = "Task" [ 867.239576] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.247904] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327744, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.283611] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.284161] env[63293]: DEBUG nova.compute.manager [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 867.287183] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.769s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.289120] env[63293]: INFO nova.compute.claims [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.418254] env[63293]: DEBUG nova.objects.base [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Object Instance<0977c79a-2ef3-4ec9-a0a0-de1ed5799723> lazy-loaded attributes: flavor,info_cache {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 867.472829] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.627421] env[63293]: DEBUG nova.compute.manager [req-caf56bd9-74bd-4b11-8803-b68dff8cc321 req-712fe4bc-eedf-4158-a50f-48806f7ffdf3 service nova] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Received event network-vif-deleted-c7cbdf4e-1a17-436f-8417-ed153aa1e71b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.627421] env[63293]: DEBUG nova.compute.manager [req-caf56bd9-74bd-4b11-8803-b68dff8cc321 req-712fe4bc-eedf-4158-a50f-48806f7ffdf3 service nova] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Received event network-vif-deleted-53a4ce4d-f5b1-46ed-b4fe-db9454cef578 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.740997] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327743, 'name': ReconfigVM_Task, 'duration_secs': 0.294687} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.745342] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Reconfigured VM instance instance-0000004a to attach disk [datastore1] d753598d-a92a-4515-9ad1-d386294f7a99/d753598d-a92a-4515-9ad1-d386294f7a99.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.745342] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1d2479d-1eb3-4c75-9496-95f5b3d01c57 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.755925] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327744, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.757802] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 867.757802] env[63293]: value = "task-1327745" [ 867.757802] env[63293]: _type = "Task" [ 867.757802] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.767511] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327745, 'name': Rename_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.796923] env[63293]: DEBUG nova.compute.utils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.798053] env[63293]: DEBUG nova.compute.manager [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 867.798216] env[63293]: DEBUG nova.network.neutron [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 868.090863] env[63293]: DEBUG nova.policy [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b80f831fedd34a6f8f438deac96cad7d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81abc1c172e44e68bee6e6fb2b74919a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 868.250055] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327744, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.814644} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.250390] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 828a1165-3829-431c-9edb-3a5d3a8054ec/828a1165-3829-431c-9edb-3a5d3a8054ec.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 868.250627] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.250891] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-744a2067-c90c-4731-a488-3aff18755376 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.257561] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 868.257561] env[63293]: value = "task-1327746" [ 868.257561] env[63293]: _type = "Task" [ 868.257561] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.268681] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327746, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.271961] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327745, 'name': Rename_Task, 'duration_secs': 0.255458} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.272294] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.272564] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46554dc5-ef78-49f1-ac86-d11ba22b2b7b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.279490] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 868.279490] env[63293]: value = "task-1327747" [ 868.279490] env[63293]: _type = "Task" [ 868.279490] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.290030] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327747, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.305044] env[63293]: DEBUG nova.compute.manager [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 868.578364] env[63293]: DEBUG nova.network.neutron [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updating instance_info_cache with network_info: [{"id": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "address": "fa:16:3e:80:11:9c", "network": {"id": "9d5a3e58-567a-4c3f-9ad9-9fcbf81fced9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-949261114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1a30d37c1b47fd8670db0a9ef7020d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f0bfb1c-61", "ovs_interfaceid": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.608132] env[63293]: DEBUG nova.network.neutron [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Successfully updated port: 50288bf0-633c-4e61-a850-e343398936c4 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 868.647687] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf8ac1a-4890-4c0f-8e95-ef4ffa53cb8d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.656818] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977864ff-e3b6-4df5-8952-c4baf0f2a4de {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.690822] env[63293]: DEBUG nova.network.neutron [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Successfully created port: fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 868.693271] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf3c123-2139-4e49-8023-a32712f2d9b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.702460] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed49398-4d34-4615-b094-8f4317ee2a71 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.718813] env[63293]: DEBUG nova.compute.provider_tree [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.768944] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327746, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072596} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.769451] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 868.770325] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6693418b-17bc-4d07-84b1-dff2ba5985b1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.791429] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 828a1165-3829-431c-9edb-3a5d3a8054ec/828a1165-3829-431c-9edb-3a5d3a8054ec.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 868.795021] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f964f2e-a2d8-4ab5-aaf0-2604a078359e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.819905] env[63293]: DEBUG oslo_vmware.api [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327747, 'name': PowerOnVM_Task, 'duration_secs': 0.474251} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.821452] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 868.821733] env[63293]: INFO nova.compute.manager [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Took 8.10 seconds to spawn the instance on the hypervisor. [ 868.821957] env[63293]: DEBUG nova.compute.manager [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.822385] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 868.822385] env[63293]: value = "task-1327748" [ 868.822385] env[63293]: _type = "Task" [ 868.822385] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.823209] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b836c18f-f7d3-4486-b29d-3f575a35f3f6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.834014] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327748, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.080914] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Releasing lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.111759] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.112099] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.112309] env[63293]: DEBUG nova.network.neutron [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 869.222629] env[63293]: DEBUG nova.scheduler.client.report [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.316950] env[63293]: DEBUG nova.compute.manager [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 869.339172] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327748, 'name': ReconfigVM_Task, 'duration_secs': 0.386059} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.343640] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 828a1165-3829-431c-9edb-3a5d3a8054ec/828a1165-3829-431c-9edb-3a5d3a8054ec.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.344422] env[63293]: INFO nova.compute.manager [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Rebuilding instance [ 869.346641] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-84a8d6e1-7094-4b15-a37d-e1bcc04fe640 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.353175] env[63293]: INFO nova.compute.manager [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Took 29.66 seconds to build instance. [ 869.358329] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 869.358329] env[63293]: value = "task-1327749" [ 869.358329] env[63293]: _type = "Task" [ 869.358329] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.367239] env[63293]: DEBUG nova.virt.hardware [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.367527] env[63293]: DEBUG nova.virt.hardware [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.367778] env[63293]: DEBUG nova.virt.hardware [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.367909] env[63293]: DEBUG nova.virt.hardware [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.368074] env[63293]: DEBUG nova.virt.hardware [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.368253] env[63293]: DEBUG nova.virt.hardware [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.368586] env[63293]: DEBUG nova.virt.hardware [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.368671] env[63293]: DEBUG nova.virt.hardware [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.368834] env[63293]: DEBUG nova.virt.hardware [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.369017] env[63293]: DEBUG nova.virt.hardware [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.369245] env[63293]: DEBUG nova.virt.hardware [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.371597] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9fa0ad-be07-46e6-8096-4dbdb4983eee {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.378377] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327749, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.391840] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88939f98-ea0a-48a0-bb13-42fc5e9248b1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.412185] env[63293]: DEBUG nova.compute.manager [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.413367] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ea23b6-4953-4437-8157-499593fc2d23 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.585590] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.586301] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf24e69c-3bad-4f88-8a94-820ca78fa817 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.594052] env[63293]: DEBUG oslo_vmware.api [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 869.594052] env[63293]: value = "task-1327750" [ 869.594052] env[63293]: _type = "Task" [ 869.594052] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.601638] env[63293]: DEBUG oslo_vmware.api [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327750, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.659270] env[63293]: WARNING nova.network.neutron [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] 908b99b3-6031-4de0-86a4-09a9cbe355a5 already exists in list: networks containing: ['908b99b3-6031-4de0-86a4-09a9cbe355a5']. ignoring it [ 869.727975] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.728609] env[63293]: DEBUG nova.compute.manager [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 869.731349] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.132s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.732979] env[63293]: INFO nova.compute.claims [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.855399] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3e405e38-2543-499c-b54d-cd3a374a504a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.193s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.869849] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327749, 'name': Rename_Task, 'duration_secs': 0.164107} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.870073] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.870308] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ebab31b8-0b83-4830-a204-86a9a3ba1091 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.878055] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 869.878055] env[63293]: value = "task-1327751" [ 869.878055] env[63293]: _type = "Task" [ 869.878055] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.886153] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327751, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.924751] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.925113] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a9b940f-99b0-4cb8-a19f-e7c5bc5bbffc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.932755] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for the task: (returnval){ [ 869.932755] env[63293]: value = "task-1327752" [ 869.932755] env[63293]: _type = "Task" [ 869.932755] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.943555] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327752, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.976356] env[63293]: DEBUG nova.network.neutron [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [{"id": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "address": "fa:16:3e:2b:1c:64", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b2cc4-af", "ovs_interfaceid": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "50288bf0-633c-4e61-a850-e343398936c4", "address": "fa:16:3e:97:22:e5", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50288bf0-63", "ovs_interfaceid": "50288bf0-633c-4e61-a850-e343398936c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.041795] env[63293]: DEBUG nova.compute.manager [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received event network-vif-plugged-50288bf0-633c-4e61-a850-e343398936c4 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.042047] env[63293]: DEBUG oslo_concurrency.lockutils [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] Acquiring lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.042288] env[63293]: DEBUG oslo_concurrency.lockutils [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.042467] env[63293]: DEBUG oslo_concurrency.lockutils [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.042729] env[63293]: DEBUG nova.compute.manager [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] No waiting events found dispatching network-vif-plugged-50288bf0-633c-4e61-a850-e343398936c4 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 870.043013] env[63293]: WARNING nova.compute.manager [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received unexpected event network-vif-plugged-50288bf0-633c-4e61-a850-e343398936c4 for instance with vm_state active and task_state None. [ 870.043302] env[63293]: DEBUG nova.compute.manager [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received event network-changed-50288bf0-633c-4e61-a850-e343398936c4 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.043580] env[63293]: DEBUG nova.compute.manager [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Refreshing instance network info cache due to event network-changed-50288bf0-633c-4e61-a850-e343398936c4. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.043835] env[63293]: DEBUG oslo_concurrency.lockutils [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] Acquiring lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.105668] env[63293]: DEBUG oslo_vmware.api [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327750, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.237265] env[63293]: DEBUG nova.compute.utils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 870.240805] env[63293]: DEBUG nova.compute.manager [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 870.240988] env[63293]: DEBUG nova.network.neutron [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 870.283480] env[63293]: DEBUG nova.policy [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc557354560049119144c62dd61d0816', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2af826d819b4cc4ab27ff8b95144039', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 870.388569] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327751, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.444914] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327752, 'name': PowerOffVM_Task, 'duration_secs': 0.1944} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.445249] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.445943] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 870.446226] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-917c3dbf-2a63-4696-8633-6e98167d8366 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.452851] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for the task: (returnval){ [ 870.452851] env[63293]: value = "task-1327753" [ 870.452851] env[63293]: _type = "Task" [ 870.452851] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.460665] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327753, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.479826] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.480623] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.480815] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.482621] env[63293]: DEBUG oslo_concurrency.lockutils [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] Acquired lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.482849] env[63293]: DEBUG nova.network.neutron [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Refreshing network info cache for port 50288bf0-633c-4e61-a850-e343398936c4 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.486028] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb70cca-1269-4c01-9cf5-0c28a145981e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.492915] env[63293]: DEBUG nova.compute.manager [req-5fbefcbe-cbb7-4795-bc97-8cbec44cb191 req-8ffbb096-1de7-4fec-9a4e-d21b84310a31 service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Received event network-vif-plugged-fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.493199] env[63293]: DEBUG oslo_concurrency.lockutils [req-5fbefcbe-cbb7-4795-bc97-8cbec44cb191 req-8ffbb096-1de7-4fec-9a4e-d21b84310a31 service nova] Acquiring lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.493728] env[63293]: DEBUG oslo_concurrency.lockutils [req-5fbefcbe-cbb7-4795-bc97-8cbec44cb191 req-8ffbb096-1de7-4fec-9a4e-d21b84310a31 service nova] Lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.493957] env[63293]: DEBUG oslo_concurrency.lockutils [req-5fbefcbe-cbb7-4795-bc97-8cbec44cb191 req-8ffbb096-1de7-4fec-9a4e-d21b84310a31 service nova] Lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.494925] env[63293]: DEBUG nova.compute.manager [req-5fbefcbe-cbb7-4795-bc97-8cbec44cb191 req-8ffbb096-1de7-4fec-9a4e-d21b84310a31 service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] No waiting events found dispatching network-vif-plugged-fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 870.498185] env[63293]: WARNING nova.compute.manager [req-5fbefcbe-cbb7-4795-bc97-8cbec44cb191 req-8ffbb096-1de7-4fec-9a4e-d21b84310a31 service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Received unexpected event network-vif-plugged-fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2 for instance with vm_state building and task_state spawning. [ 870.519132] env[63293]: DEBUG nova.virt.hardware [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.519557] env[63293]: DEBUG nova.virt.hardware [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.519801] env[63293]: DEBUG nova.virt.hardware [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.520050] env[63293]: DEBUG nova.virt.hardware [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.520255] env[63293]: DEBUG nova.virt.hardware [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.520479] env[63293]: DEBUG nova.virt.hardware [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.520943] env[63293]: DEBUG nova.virt.hardware [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.520943] env[63293]: DEBUG nova.virt.hardware [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.521463] env[63293]: DEBUG nova.virt.hardware [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.521463] env[63293]: DEBUG nova.virt.hardware [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.521616] env[63293]: DEBUG nova.virt.hardware [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.528936] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Reconfiguring VM to attach interface {{(pid=63293) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 870.529888] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d6e0cb9-2601-458c-8d72-7980d641f4ec {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.554436] env[63293]: DEBUG oslo_vmware.api [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 870.554436] env[63293]: value = "task-1327754" [ 870.554436] env[63293]: _type = "Task" [ 870.554436] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.555241] env[63293]: DEBUG nova.network.neutron [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Successfully updated port: fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 870.567672] env[63293]: DEBUG oslo_vmware.api [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327754, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.568708] env[63293]: DEBUG nova.network.neutron [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Successfully created port: 5533b032-6f43-4c5e-92e7-da17c807dbc1 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 870.605506] env[63293]: DEBUG oslo_vmware.api [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327750, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.745023] env[63293]: DEBUG nova.compute.manager [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 870.892525] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327751, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.965453] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] VM already powered off {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 870.965700] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Volume detach. Driver type: vmdk {{(pid=63293) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 870.965896] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283694', 'volume_id': '1f5a039e-f604-47d3-a580-f1754a5b697d', 'name': 'volume-1f5a039e-f604-47d3-a580-f1754a5b697d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8637881-9fb6-4c5d-848d-7b2d38f8e970', 'attached_at': '', 'detached_at': '', 'volume_id': '1f5a039e-f604-47d3-a580-f1754a5b697d', 'serial': '1f5a039e-f604-47d3-a580-f1754a5b697d'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 870.966717] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3db322-10d5-4cf1-8aae-b02254b31738 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.988604] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b58685a-32aa-4fc9-80df-dbacf88c72eb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.997413] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e98534f-1e38-4543-8fcb-0afc7ce2b6a8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.019287] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e07ff8e-1c39-4cf2-8499-370d56fc69f9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.036435] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] The volume has not been displaced from its original location: [datastore2] volume-1f5a039e-f604-47d3-a580-f1754a5b697d/volume-1f5a039e-f604-47d3-a580-f1754a5b697d.vmdk. No consolidation needed. {{(pid=63293) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 871.043106] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Reconfiguring VM instance instance-0000003d to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 871.045876] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bea869f-3a35-4e6e-ab33-848f47495870 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.064139] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquiring lock "refresh_cache-a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.064332] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquired lock "refresh_cache-a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.064575] env[63293]: DEBUG nova.network.neutron [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.075863] env[63293]: DEBUG oslo_vmware.api [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327754, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.077980] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for the task: (returnval){ [ 871.077980] env[63293]: value = "task-1327755" [ 871.077980] env[63293]: _type = "Task" [ 871.077980] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.089324] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327755, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.107490] env[63293]: DEBUG oslo_vmware.api [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327750, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.145983] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13e3165-5b71-47b7-aa06-2a26b4969c45 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.152922] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e6f338-ce38-478e-a142-ab5dbb9a6f98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.186655] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794679bb-1d7f-4d95-8098-8cebe380d8eb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.195038] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11dcf6ca-87aa-4c0a-8693-ef8b7a79f57a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.211838] env[63293]: DEBUG nova.compute.provider_tree [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.353329] env[63293]: DEBUG nova.network.neutron [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updated VIF entry in instance network info cache for port 50288bf0-633c-4e61-a850-e343398936c4. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.353947] env[63293]: DEBUG nova.network.neutron [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [{"id": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "address": "fa:16:3e:2b:1c:64", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b2cc4-af", "ovs_interfaceid": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "50288bf0-633c-4e61-a850-e343398936c4", "address": "fa:16:3e:97:22:e5", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50288bf0-63", "ovs_interfaceid": "50288bf0-633c-4e61-a850-e343398936c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.391267] env[63293]: DEBUG oslo_vmware.api [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327751, 'name': PowerOnVM_Task, 'duration_secs': 1.257186} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.391601] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 871.391762] env[63293]: INFO nova.compute.manager [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Took 6.36 seconds to spawn the instance on the hypervisor. [ 871.391941] env[63293]: DEBUG nova.compute.manager [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 871.393257] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9a9df5-29d6-4ea7-8754-897b195d67ef {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.577884] env[63293]: DEBUG oslo_vmware.api [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327754, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.587464] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327755, 'name': ReconfigVM_Task, 'duration_secs': 0.320042} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.587788] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Reconfigured VM instance instance-0000003d to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 871.592717] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cbc4b3d-525c-469d-84f8-968f6f711ed3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.611630] env[63293]: DEBUG oslo_vmware.api [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327750, 'name': PowerOnVM_Task, 'duration_secs': 1.842231} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.612859] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 871.613085] env[63293]: DEBUG nova.compute.manager [None req-8d0cfbac-e877-4828-9684-5c576457bd7a tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 871.613432] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for the task: (returnval){ [ 871.613432] env[63293]: value = "task-1327756" [ 871.613432] env[63293]: _type = "Task" [ 871.613432] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.614151] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970d54a9-d968-4a5e-ace0-8ebb5faeeb98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.628733] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327756, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.641706] env[63293]: DEBUG nova.network.neutron [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.715368] env[63293]: DEBUG nova.scheduler.client.report [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.756326] env[63293]: DEBUG nova.compute.manager [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 871.789882] env[63293]: DEBUG nova.virt.hardware [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 871.790186] env[63293]: DEBUG nova.virt.hardware [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 871.790310] env[63293]: DEBUG nova.virt.hardware [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 871.790514] env[63293]: DEBUG nova.virt.hardware [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 871.790663] env[63293]: DEBUG nova.virt.hardware [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 871.790812] env[63293]: DEBUG nova.virt.hardware [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 871.791038] env[63293]: DEBUG nova.virt.hardware [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 871.791207] env[63293]: DEBUG nova.virt.hardware [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 871.791415] env[63293]: DEBUG nova.virt.hardware [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 871.791630] env[63293]: DEBUG nova.virt.hardware [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 871.791918] env[63293]: DEBUG nova.virt.hardware [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 871.793278] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647aa459-1963-4039-b7f7-22f017ea6479 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.805544] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b56be4f-e07b-4e56-b1b8-fe3c71dd5430 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.855494] env[63293]: DEBUG nova.network.neutron [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Updating instance_info_cache with network_info: [{"id": "fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2", "address": "fa:16:3e:5f:0f:cc", "network": {"id": "9d41bc05-b8ef-407d-b05d-cc50a54820a5", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-447925489-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81abc1c172e44e68bee6e6fb2b74919a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfddcef04-2e", "ovs_interfaceid": "fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.856657] env[63293]: DEBUG oslo_concurrency.lockutils [req-40e8c266-b8fb-4fe9-b8b1-293c2f013921 req-490210e2-9e85-49cb-be67-29214a4288b2 service nova] Releasing lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.911041] env[63293]: INFO nova.compute.manager [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Took 28.24 seconds to build instance. [ 872.076026] env[63293]: DEBUG oslo_vmware.api [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327754, 'name': ReconfigVM_Task, 'duration_secs': 1.081564} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.076163] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.076407] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Reconfigured VM to attach interface {{(pid=63293) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 872.135352] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327756, 'name': ReconfigVM_Task, 'duration_secs': 0.15833} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.137023] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283694', 'volume_id': '1f5a039e-f604-47d3-a580-f1754a5b697d', 'name': 'volume-1f5a039e-f604-47d3-a580-f1754a5b697d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8637881-9fb6-4c5d-848d-7b2d38f8e970', 'attached_at': '', 'detached_at': '', 'volume_id': '1f5a039e-f604-47d3-a580-f1754a5b697d', 'serial': '1f5a039e-f604-47d3-a580-f1754a5b697d'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 872.137277] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.139180] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20e5fbc-91a9-4e0f-86eb-1ea9c61a2aa7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.148319] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.148664] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-afe0b735-f4bb-4134-b82b-e45722c8f04a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.222489] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.223055] env[63293]: DEBUG nova.compute.manager [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 872.225934] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.652s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.227346] env[63293]: INFO nova.compute.claims [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.257053] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.257053] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.257053] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Deleting the datastore file [datastore2] b8637881-9fb6-4c5d-848d-7b2d38f8e970 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.257053] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-204b21d4-912a-480f-9e05-f7faabaea47b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.263751] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for the task: (returnval){ [ 872.263751] env[63293]: value = "task-1327758" [ 872.263751] env[63293]: _type = "Task" [ 872.263751] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.272953] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327758, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.358651] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Releasing lock "refresh_cache-a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.358986] env[63293]: DEBUG nova.compute.manager [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Instance network_info: |[{"id": "fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2", "address": "fa:16:3e:5f:0f:cc", "network": {"id": "9d41bc05-b8ef-407d-b05d-cc50a54820a5", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-447925489-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81abc1c172e44e68bee6e6fb2b74919a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfddcef04-2e", "ovs_interfaceid": "fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 872.359533] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:0f:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd96b39f-bd2e-48d1-85c3-577cf97f08c8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.367036] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Creating folder: Project (81abc1c172e44e68bee6e6fb2b74919a). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.367366] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-efda0c79-b599-4ea8-ab12-1f796a504cb2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.378200] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Created folder: Project (81abc1c172e44e68bee6e6fb2b74919a) in parent group-v283678. [ 872.378411] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Creating folder: Instances. Parent ref: group-v283772. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.378661] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-577aa8e1-9a59-425c-b3dd-d6a8a48f4de6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.386948] env[63293]: DEBUG nova.compute.manager [req-c42f1369-c114-42ea-bc32-dc28cf0b2b91 req-43d06b36-5be9-4011-8267-c3f66d146e14 service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Received event network-vif-plugged-5533b032-6f43-4c5e-92e7-da17c807dbc1 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.387213] env[63293]: DEBUG oslo_concurrency.lockutils [req-c42f1369-c114-42ea-bc32-dc28cf0b2b91 req-43d06b36-5be9-4011-8267-c3f66d146e14 service nova] Acquiring lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.387488] env[63293]: DEBUG oslo_concurrency.lockutils [req-c42f1369-c114-42ea-bc32-dc28cf0b2b91 req-43d06b36-5be9-4011-8267-c3f66d146e14 service nova] Lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.387602] env[63293]: DEBUG oslo_concurrency.lockutils [req-c42f1369-c114-42ea-bc32-dc28cf0b2b91 req-43d06b36-5be9-4011-8267-c3f66d146e14 service nova] Lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.387777] env[63293]: DEBUG nova.compute.manager [req-c42f1369-c114-42ea-bc32-dc28cf0b2b91 req-43d06b36-5be9-4011-8267-c3f66d146e14 service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] No waiting events found dispatching network-vif-plugged-5533b032-6f43-4c5e-92e7-da17c807dbc1 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 872.387967] env[63293]: WARNING nova.compute.manager [req-c42f1369-c114-42ea-bc32-dc28cf0b2b91 req-43d06b36-5be9-4011-8267-c3f66d146e14 service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Received unexpected event network-vif-plugged-5533b032-6f43-4c5e-92e7-da17c807dbc1 for instance with vm_state building and task_state spawning. [ 872.389911] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Created folder: Instances in parent group-v283772. [ 872.390191] env[63293]: DEBUG oslo.service.loopingcall [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.390390] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.390631] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e308cd1a-9bf7-48a2-b8e3-e0a557eda326 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.411977] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.411977] env[63293]: value = "task-1327761" [ 872.411977] env[63293]: _type = "Task" [ 872.411977] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.415189] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b76d0c09-231a-4cdc-bdd2-4ca253cf8f62 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "828a1165-3829-431c-9edb-3a5d3a8054ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.333s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.420828] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327761, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.582072] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d43294-8a54-4ac5-a389-fde2256de60e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.295s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.628036] env[63293]: INFO nova.compute.manager [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Rebuilding instance [ 872.655069] env[63293]: DEBUG nova.network.neutron [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Successfully updated port: 5533b032-6f43-4c5e-92e7-da17c807dbc1 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 872.678660] env[63293]: DEBUG nova.compute.manager [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.680081] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908b3082-544a-4096-a86b-cff516089a92 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.731968] env[63293]: DEBUG nova.compute.utils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.736396] env[63293]: DEBUG nova.compute.manager [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.736607] env[63293]: DEBUG nova.network.neutron [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 872.775328] env[63293]: DEBUG oslo_vmware.api [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Task: {'id': task-1327758, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123052} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.775810] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.776149] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 872.776449] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 872.794854] env[63293]: DEBUG nova.policy [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6860a9e34a1b44029cf57d9f04ae87c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fac34e49d5dc49e7a7055c998d5b6766', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.802622] env[63293]: DEBUG nova.compute.manager [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Received event network-changed-fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.802847] env[63293]: DEBUG nova.compute.manager [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Refreshing instance network info cache due to event network-changed-fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 872.803404] env[63293]: DEBUG oslo_concurrency.lockutils [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] Acquiring lock "refresh_cache-a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.803739] env[63293]: DEBUG oslo_concurrency.lockutils [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] Acquired lock "refresh_cache-a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.803798] env[63293]: DEBUG nova.network.neutron [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Refreshing network info cache for port fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.848161] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Volume detach. Driver type: vmdk {{(pid=63293) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 872.848806] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6caebf47-32d5-435a-932e-5d3ce6088dd4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.859919] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45680779-01f9-4249-9126-3b8be44a5145 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.874541] env[63293]: DEBUG nova.compute.manager [req-68c2e694-58e3-449e-8bec-a396d0f1f618 req-0e249f29-61a9-470e-ba02-4c9ca548249a service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Received event network-changed-5533b032-6f43-4c5e-92e7-da17c807dbc1 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.874813] env[63293]: DEBUG nova.compute.manager [req-68c2e694-58e3-449e-8bec-a396d0f1f618 req-0e249f29-61a9-470e-ba02-4c9ca548249a service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Refreshing instance network info cache due to event network-changed-5533b032-6f43-4c5e-92e7-da17c807dbc1. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 872.875025] env[63293]: DEBUG oslo_concurrency.lockutils [req-68c2e694-58e3-449e-8bec-a396d0f1f618 req-0e249f29-61a9-470e-ba02-4c9ca548249a service nova] Acquiring lock "refresh_cache-5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.875118] env[63293]: DEBUG oslo_concurrency.lockutils [req-68c2e694-58e3-449e-8bec-a396d0f1f618 req-0e249f29-61a9-470e-ba02-4c9ca548249a service nova] Acquired lock "refresh_cache-5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.875304] env[63293]: DEBUG nova.network.neutron [req-68c2e694-58e3-449e-8bec-a396d0f1f618 req-0e249f29-61a9-470e-ba02-4c9ca548249a service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Refreshing network info cache for port 5533b032-6f43-4c5e-92e7-da17c807dbc1 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.891007] env[63293]: ERROR nova.compute.manager [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Failed to detach volume 1f5a039e-f604-47d3-a580-f1754a5b697d from /dev/sda: nova.exception.InstanceNotFound: Instance b8637881-9fb6-4c5d-848d-7b2d38f8e970 could not be found. [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Traceback (most recent call last): [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self.driver.rebuild(**kwargs) [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] raise NotImplementedError() [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] NotImplementedError [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] During handling of the above exception, another exception occurred: [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Traceback (most recent call last): [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self.driver.detach_volume(context, old_connection_info, [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] return self._volumeops.detach_volume(connection_info, instance) [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self._detach_volume_vmdk(connection_info, instance) [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] stable_ref.fetch_moref(session) [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] raise exception.InstanceNotFound(instance_id=self._uuid) [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] nova.exception.InstanceNotFound: Instance b8637881-9fb6-4c5d-848d-7b2d38f8e970 could not be found. [ 872.891007] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] [ 872.927736] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327761, 'name': CreateVM_Task, 'duration_secs': 0.415496} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.927926] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 872.929751] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.929751] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.929751] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.929751] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5dc730d3-27cc-4ea5-ac80-3770b70df67f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.935464] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for the task: (returnval){ [ 872.935464] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52752771-521b-7490-6afa-812e1d1b30b3" [ 872.935464] env[63293]: _type = "Task" [ 872.935464] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.943937] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52752771-521b-7490-6afa-812e1d1b30b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.052034] env[63293]: DEBUG nova.compute.utils [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Build of instance b8637881-9fb6-4c5d-848d-7b2d38f8e970 aborted: Failed to rebuild volume backed instance. {{(pid=63293) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 873.054377] env[63293]: ERROR nova.compute.manager [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance b8637881-9fb6-4c5d-848d-7b2d38f8e970 aborted: Failed to rebuild volume backed instance. [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Traceback (most recent call last): [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self.driver.rebuild(**kwargs) [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] raise NotImplementedError() [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] NotImplementedError [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] During handling of the above exception, another exception occurred: [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Traceback (most recent call last): [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self._detach_root_volume(context, instance, root_bdm) [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] with excutils.save_and_reraise_exception(): [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self.force_reraise() [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] raise self.value [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self.driver.detach_volume(context, old_connection_info, [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] return self._volumeops.detach_volume(connection_info, instance) [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self._detach_volume_vmdk(connection_info, instance) [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] stable_ref.fetch_moref(session) [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] raise exception.InstanceNotFound(instance_id=self._uuid) [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] nova.exception.InstanceNotFound: Instance b8637881-9fb6-4c5d-848d-7b2d38f8e970 could not be found. [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] During handling of the above exception, another exception occurred: [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Traceback (most recent call last): [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] yield [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 873.054377] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self._do_rebuild_instance_with_claim( [ 873.056218] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 873.056218] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self._do_rebuild_instance( [ 873.056218] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 873.056218] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self._rebuild_default_impl(**kwargs) [ 873.056218] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 873.056218] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] self._rebuild_volume_backed_instance( [ 873.056218] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 873.056218] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] raise exception.BuildAbortException( [ 873.056218] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] nova.exception.BuildAbortException: Build of instance b8637881-9fb6-4c5d-848d-7b2d38f8e970 aborted: Failed to rebuild volume backed instance. [ 873.056218] env[63293]: ERROR nova.compute.manager [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] [ 873.071652] env[63293]: DEBUG nova.network.neutron [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Successfully created port: 46d5563f-41a2-48ab-9d32-3aa394a13da8 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.158881] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "refresh_cache-5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.193838] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.194205] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98b3dc39-7bbc-4dad-a938-a3511e025dae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.204263] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 873.204263] env[63293]: value = "task-1327762" [ 873.204263] env[63293]: _type = "Task" [ 873.204263] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.211960] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327762, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.240724] env[63293]: DEBUG nova.compute.manager [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.438128] env[63293]: DEBUG nova.network.neutron [req-68c2e694-58e3-449e-8bec-a396d0f1f618 req-0e249f29-61a9-470e-ba02-4c9ca548249a service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 873.461463] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52752771-521b-7490-6afa-812e1d1b30b3, 'name': SearchDatastore_Task, 'duration_secs': 0.015705} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.462136] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.462573] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.462932] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.463223] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.463583] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.464040] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44a0954c-dfb3-436d-981b-4b7aaf9896b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.479364] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.479364] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 873.479364] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5fd4cda-76cc-4b93-b62f-31487437f90d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.490152] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for the task: (returnval){ [ 873.490152] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52db8456-c2ac-f588-2ce0-67d87284e5cc" [ 873.490152] env[63293]: _type = "Task" [ 873.490152] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.502160] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52db8456-c2ac-f588-2ce0-67d87284e5cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.586717] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670670b7-f95e-459b-8455-40fff9b528c0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.597825] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6872a36f-8e8b-4d4d-b9b8-3a06e67eb9a0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.631789] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f63e43-b5a0-493f-96c3-4c0178ac5236 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.639906] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3619e7bf-f0cd-4fc9-8f02-c893939d9b55 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.644832] env[63293]: DEBUG nova.network.neutron [req-68c2e694-58e3-449e-8bec-a396d0f1f618 req-0e249f29-61a9-470e-ba02-4c9ca548249a service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.657498] env[63293]: DEBUG nova.compute.provider_tree [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.714262] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327762, 'name': PowerOffVM_Task, 'duration_secs': 0.205162} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.714571] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.715973] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.715973] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe83960d-9ff2-48f1-97c4-6aff8ada64a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.723902] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.723902] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90867888-b719-4eba-9c09-2f272e5cd1d7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.737087] env[63293]: DEBUG nova.network.neutron [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Updated VIF entry in instance network info cache for port fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 873.737356] env[63293]: DEBUG nova.network.neutron [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Updating instance_info_cache with network_info: [{"id": "fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2", "address": "fa:16:3e:5f:0f:cc", "network": {"id": "9d41bc05-b8ef-407d-b05d-cc50a54820a5", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-447925489-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81abc1c172e44e68bee6e6fb2b74919a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfddcef04-2e", "ovs_interfaceid": "fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.750943] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.750943] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.751122] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Deleting the datastore file [datastore1] 828a1165-3829-431c-9edb-3a5d3a8054ec {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.751641] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-778f354b-29e3-4817-bc0a-72bef19c60e1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.758179] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 873.758179] env[63293]: value = "task-1327764" [ 873.758179] env[63293]: _type = "Task" [ 873.758179] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.766823] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327764, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.869474] env[63293]: DEBUG oslo_vmware.rw_handles [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d134c-1e3f-e738-6d20-c536c1bd38a7/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 873.870455] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7cfad5-bf99-4e17-afdd-837b9524c409 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.876938] env[63293]: DEBUG oslo_vmware.rw_handles [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d134c-1e3f-e738-6d20-c536c1bd38a7/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 873.876938] env[63293]: ERROR oslo_vmware.rw_handles [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d134c-1e3f-e738-6d20-c536c1bd38a7/disk-0.vmdk due to incomplete transfer. [ 873.876938] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-60e500c8-0148-467c-b13f-12371f79c93d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.882910] env[63293]: DEBUG oslo_vmware.rw_handles [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d134c-1e3f-e738-6d20-c536c1bd38a7/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 873.883124] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Uploaded image 46aa5632-364b-4607-b217-b473d1ef29ad to the Glance image server {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 873.885397] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Destroying the VM {{(pid=63293) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 873.885652] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-09e6e444-d4c9-44e0-a1c9-826858234800 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.891331] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 873.891331] env[63293]: value = "task-1327765" [ 873.891331] env[63293]: _type = "Task" [ 873.891331] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.899404] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327765, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.008915] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52db8456-c2ac-f588-2ce0-67d87284e5cc, 'name': SearchDatastore_Task, 'duration_secs': 0.011601} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.011160] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-904c8515-59ba-42a8-a439-5161f24d192e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.020747] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for the task: (returnval){ [ 874.020747] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5214f459-6863-84e2-ca7b-0216629185fd" [ 874.020747] env[63293]: _type = "Task" [ 874.020747] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.029688] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5214f459-6863-84e2-ca7b-0216629185fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.147390] env[63293]: DEBUG oslo_concurrency.lockutils [req-68c2e694-58e3-449e-8bec-a396d0f1f618 req-0e249f29-61a9-470e-ba02-4c9ca548249a service nova] Releasing lock "refresh_cache-5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.147768] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "refresh_cache-5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.147982] env[63293]: DEBUG nova.network.neutron [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.160977] env[63293]: DEBUG nova.scheduler.client.report [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.243897] env[63293]: DEBUG oslo_concurrency.lockutils [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] Releasing lock "refresh_cache-a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.244276] env[63293]: DEBUG nova.compute.manager [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Received event network-changed-d41cf464-2396-4e22-80d1-3c1602939528 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.244525] env[63293]: DEBUG nova.compute.manager [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Refreshing instance network info cache due to event network-changed-d41cf464-2396-4e22-80d1-3c1602939528. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 874.244573] env[63293]: DEBUG oslo_concurrency.lockutils [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] Acquiring lock "refresh_cache-d753598d-a92a-4515-9ad1-d386294f7a99" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.244699] env[63293]: DEBUG oslo_concurrency.lockutils [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] Acquired lock "refresh_cache-d753598d-a92a-4515-9ad1-d386294f7a99" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.244891] env[63293]: DEBUG nova.network.neutron [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Refreshing network info cache for port d41cf464-2396-4e22-80d1-3c1602939528 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 874.252736] env[63293]: DEBUG nova.compute.manager [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 874.269884] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327764, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104217} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.270156] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.270374] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 874.270558] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.281720] env[63293]: DEBUG nova.virt.hardware [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.281947] env[63293]: DEBUG nova.virt.hardware [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.282113] env[63293]: DEBUG nova.virt.hardware [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.283025] env[63293]: DEBUG nova.virt.hardware [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.283025] env[63293]: DEBUG nova.virt.hardware [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.283025] env[63293]: DEBUG nova.virt.hardware [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.283025] env[63293]: DEBUG nova.virt.hardware [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.283263] env[63293]: DEBUG nova.virt.hardware [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.283263] env[63293]: DEBUG nova.virt.hardware [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.283437] env[63293]: DEBUG nova.virt.hardware [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.283592] env[63293]: DEBUG nova.virt.hardware [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.284468] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e79368-8c8d-4542-ba0e-a6a3e793018e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.293000] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89983eb0-b458-46e3-ada2-140af1e27b66 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.401497] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327765, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.531273] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5214f459-6863-84e2-ca7b-0216629185fd, 'name': SearchDatastore_Task, 'duration_secs': 0.00956} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.531813] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.531908] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] a15fc833-d9e8-4340-ab4f-c4a0b2d0e703/a15fc833-d9e8-4340-ab4f-c4a0b2d0e703.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 874.532122] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-beddd2fa-b5d5-47c7-b487-2354dd6eb15d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.538701] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for the task: (returnval){ [ 874.538701] env[63293]: value = "task-1327766" [ 874.538701] env[63293]: _type = "Task" [ 874.538701] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.546557] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327766, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.666248] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.666864] env[63293]: DEBUG nova.compute.manager [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 874.669814] env[63293]: DEBUG oslo_concurrency.lockutils [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.901s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.670040] env[63293]: DEBUG nova.objects.instance [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lazy-loading 'resources' on Instance uuid c26d6a3a-3643-4b52-870a-4e2ed901d908 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.682926] env[63293]: DEBUG nova.network.neutron [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.846177] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-e777ae42-3549-4573-866c-8353417608f1" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.846177] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-e777ae42-3549-4573-866c-8353417608f1" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.846657] env[63293]: DEBUG nova.objects.instance [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'flavor' on Instance uuid 65c2f72d-6bbe-45ac-8efc-401e0664390f {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.903684] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327765, 'name': Destroy_Task, 'duration_secs': 0.963485} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.903977] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Destroyed the VM [ 874.904310] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Deleting Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 874.904651] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1bfc05e1-e23b-4887-8273-2144e1e43791 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.911438] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 874.911438] env[63293]: value = "task-1327767" [ 874.911438] env[63293]: _type = "Task" [ 874.911438] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.932129] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327767, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.933323] env[63293]: DEBUG nova.network.neutron [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Updating instance_info_cache with network_info: [{"id": "5533b032-6f43-4c5e-92e7-da17c807dbc1", "address": "fa:16:3e:1c:93:e3", "network": {"id": "340bfd3b-7dff-454b-87df-788003988adc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-388838680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2af826d819b4cc4ab27ff8b95144039", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5533b032-6f", "ovs_interfaceid": "5533b032-6f43-4c5e-92e7-da17c807dbc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.044138] env[63293]: DEBUG nova.network.neutron [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Successfully updated port: 46d5563f-41a2-48ab-9d32-3aa394a13da8 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.050747] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327766, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505797} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.051377] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] a15fc833-d9e8-4340-ab4f-c4a0b2d0e703/a15fc833-d9e8-4340-ab4f-c4a0b2d0e703.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.051377] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.051825] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6757ba12-db5c-4925-ba22-83fcaa5ddee7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.058103] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for the task: (returnval){ [ 875.058103] env[63293]: value = "task-1327768" [ 875.058103] env[63293]: _type = "Task" [ 875.058103] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.066755] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327768, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.068537] env[63293]: DEBUG nova.network.neutron [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Updated VIF entry in instance network info cache for port d41cf464-2396-4e22-80d1-3c1602939528. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 875.068864] env[63293]: DEBUG nova.network.neutron [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Updating instance_info_cache with network_info: [{"id": "d41cf464-2396-4e22-80d1-3c1602939528", "address": "fa:16:3e:1e:ad:a7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd41cf464-23", "ovs_interfaceid": "d41cf464-2396-4e22-80d1-3c1602939528", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.077701] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.173142] env[63293]: DEBUG nova.compute.utils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 875.177925] env[63293]: DEBUG nova.compute.manager [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 875.178087] env[63293]: DEBUG nova.network.neutron [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 875.232398] env[63293]: DEBUG nova.policy [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4098c09e1c994cbbbb578fa6a3a5a002', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1cf310a9134147caa91f6aec4ebde76e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 875.309371] env[63293]: DEBUG nova.virt.hardware [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.309676] env[63293]: DEBUG nova.virt.hardware [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.309837] env[63293]: DEBUG nova.virt.hardware [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.310079] env[63293]: DEBUG nova.virt.hardware [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.310283] env[63293]: DEBUG nova.virt.hardware [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.310398] env[63293]: DEBUG nova.virt.hardware [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.311038] env[63293]: DEBUG nova.virt.hardware [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.311038] env[63293]: DEBUG nova.virt.hardware [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.311038] env[63293]: DEBUG nova.virt.hardware [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.311185] env[63293]: DEBUG nova.virt.hardware [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.311340] env[63293]: DEBUG nova.virt.hardware [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.312201] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300ee172-5784-49c7-b2ec-2185a44c2092 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.322829] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d85324a-34bf-4309-84ae-dad9d08be1a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.338971] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.344595] env[63293]: DEBUG oslo.service.loopingcall [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.347121] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.348025] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d7301319-35d9-4e2f-bf92-a3cd629f0f12 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.370337] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.370337] env[63293]: value = "task-1327769" [ 875.370337] env[63293]: _type = "Task" [ 875.370337] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.377752] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327769, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.420925] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327767, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.438668] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "refresh_cache-5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.438984] env[63293]: DEBUG nova.compute.manager [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Instance network_info: |[{"id": "5533b032-6f43-4c5e-92e7-da17c807dbc1", "address": "fa:16:3e:1c:93:e3", "network": {"id": "340bfd3b-7dff-454b-87df-788003988adc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-388838680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2af826d819b4cc4ab27ff8b95144039", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5533b032-6f", "ovs_interfaceid": "5533b032-6f43-4c5e-92e7-da17c807dbc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.439485] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:93:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb2eaecd-9701-4504-9fcb-fb1a420ead72', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5533b032-6f43-4c5e-92e7-da17c807dbc1', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.447476] env[63293]: DEBUG oslo.service.loopingcall [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.450695] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.451945] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c39ec178-6e0a-4914-a80e-39d21e14af42 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.468458] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7331e00b-1bd3-4928-a512-57f35ba9eb74 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.477257] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3536a9-365a-453a-be6e-df60d8b94216 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.480387] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.480387] env[63293]: value = "task-1327770" [ 875.480387] env[63293]: _type = "Task" [ 875.480387] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.481370] env[63293]: DEBUG nova.objects.instance [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'pci_requests' on Instance uuid 65c2f72d-6bbe-45ac-8efc-401e0664390f {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.516754] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9848524-613c-48a3-9e59-dc9645f4b121 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.520770] env[63293]: DEBUG nova.compute.manager [req-d550a2d5-e6cc-447f-9b9f-7f068cc57703 req-a3439308-5ce4-4b4d-8234-2ea9d761232f service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Received event network-vif-plugged-46d5563f-41a2-48ab-9d32-3aa394a13da8 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.520919] env[63293]: DEBUG oslo_concurrency.lockutils [req-d550a2d5-e6cc-447f-9b9f-7f068cc57703 req-a3439308-5ce4-4b4d-8234-2ea9d761232f service nova] Acquiring lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.521174] env[63293]: DEBUG oslo_concurrency.lockutils [req-d550a2d5-e6cc-447f-9b9f-7f068cc57703 req-a3439308-5ce4-4b4d-8234-2ea9d761232f service nova] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.521352] env[63293]: DEBUG oslo_concurrency.lockutils [req-d550a2d5-e6cc-447f-9b9f-7f068cc57703 req-a3439308-5ce4-4b4d-8234-2ea9d761232f service nova] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.521518] env[63293]: DEBUG nova.compute.manager [req-d550a2d5-e6cc-447f-9b9f-7f068cc57703 req-a3439308-5ce4-4b4d-8234-2ea9d761232f service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] No waiting events found dispatching network-vif-plugged-46d5563f-41a2-48ab-9d32-3aa394a13da8 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.521687] env[63293]: WARNING nova.compute.manager [req-d550a2d5-e6cc-447f-9b9f-7f068cc57703 req-a3439308-5ce4-4b4d-8234-2ea9d761232f service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Received unexpected event network-vif-plugged-46d5563f-41a2-48ab-9d32-3aa394a13da8 for instance with vm_state building and task_state spawning. [ 875.525858] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327770, 'name': CreateVM_Task} progress is 15%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.531720] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581c22c1-9ca8-4caf-8d5e-6c4fcf66c9f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.548291] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.548291] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.548291] env[63293]: DEBUG nova.network.neutron [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.548829] env[63293]: DEBUG nova.compute.provider_tree [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.569105] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327768, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.132593} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.569410] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.570557] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c4a99e-00ff-4f99-9763-0af73d505efa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.573539] env[63293]: DEBUG nova.network.neutron [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Successfully created port: dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.575453] env[63293]: DEBUG oslo_concurrency.lockutils [req-0c48decc-f4e4-4f07-b869-75259d62454f req-b4eb6d80-1e25-4aec-b87e-88d4aaa3d2bf service nova] Releasing lock "refresh_cache-d753598d-a92a-4515-9ad1-d386294f7a99" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.596119] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] a15fc833-d9e8-4340-ab4f-c4a0b2d0e703/a15fc833-d9e8-4340-ab4f-c4a0b2d0e703.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.596965] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4cb0d88-8bcc-4ea8-b9af-d8507849a256 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.617489] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for the task: (returnval){ [ 875.617489] env[63293]: value = "task-1327771" [ 875.617489] env[63293]: _type = "Task" [ 875.617489] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.627866] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327771, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.678984] env[63293]: DEBUG nova.compute.manager [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 875.881595] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327769, 'name': CreateVM_Task, 'duration_secs': 0.490965} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.881792] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.882224] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.882383] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.882724] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.882981] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57e240b0-dc2f-475a-80a1-283e2e7f55b3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.887630] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 875.887630] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f065a5-f40c-0cc9-aa25-4aa8ff55fadc" [ 875.887630] env[63293]: _type = "Task" [ 875.887630] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.895639] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f065a5-f40c-0cc9-aa25-4aa8ff55fadc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.920970] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327767, 'name': RemoveSnapshot_Task, 'duration_secs': 1.004519} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.921257] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Deleted Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 875.921529] env[63293]: DEBUG nova.compute.manager [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 875.922315] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce3467e3-564d-447b-981e-587ab9423d47 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.992192] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327770, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.013415] env[63293]: DEBUG nova.objects.base [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Object Instance<65c2f72d-6bbe-45ac-8efc-401e0664390f> lazy-loaded attributes: flavor,pci_requests {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 876.013635] env[63293]: DEBUG nova.network.neutron [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 876.052878] env[63293]: DEBUG nova.scheduler.client.report [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.082169] env[63293]: DEBUG nova.network.neutron [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 876.110898] env[63293]: DEBUG nova.policy [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2312d48038d7474481a0709267645780', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '866b347100454019b07f63922b995bb6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 876.127270] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327771, 'name': ReconfigVM_Task, 'duration_secs': 0.30735} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.127549] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Reconfigured VM instance instance-0000004c to attach disk [datastore2] a15fc833-d9e8-4340-ab4f-c4a0b2d0e703/a15fc833-d9e8-4340-ab4f-c4a0b2d0e703.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.128198] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a789bea-4588-45b1-8cf1-48f36618d053 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.144164] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for the task: (returnval){ [ 876.144164] env[63293]: value = "task-1327772" [ 876.144164] env[63293]: _type = "Task" [ 876.144164] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.158685] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327772, 'name': Rename_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.241815] env[63293]: DEBUG nova.network.neutron [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance_info_cache with network_info: [{"id": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "address": "fa:16:3e:12:25:eb", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46d5563f-41", "ovs_interfaceid": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.398499] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f065a5-f40c-0cc9-aa25-4aa8ff55fadc, 'name': SearchDatastore_Task, 'duration_secs': 0.012377} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.398882] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.399103] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.399376] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.399532] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.399720] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.400425] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e36b81b2-0060-42ce-b8f5-36386cb8096e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.408076] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.408258] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 876.408945] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3de3b616-4025-4416-83e3-1ef5840447d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.413886] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 876.413886] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a3bb00-f9aa-95bd-3625-cac7d00138fc" [ 876.413886] env[63293]: _type = "Task" [ 876.413886] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.421952] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a3bb00-f9aa-95bd-3625-cac7d00138fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.433837] env[63293]: INFO nova.compute.manager [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Shelve offloading [ 876.435388] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.435627] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-febe4c2b-d60b-4c0a-9c90-7f3872327a56 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.442797] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 876.442797] env[63293]: value = "task-1327773" [ 876.442797] env[63293]: _type = "Task" [ 876.442797] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.450848] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327773, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.493335] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327770, 'name': CreateVM_Task, 'duration_secs': 0.561773} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.493473] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 876.494165] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.494340] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.494664] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.494909] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bc880f5-30f3-4462-b5ed-5bdc1eb13162 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.500100] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 876.500100] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524aab2e-bfa0-63d3-a0c5-c59fb6af57b9" [ 876.500100] env[63293]: _type = "Task" [ 876.500100] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.508321] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524aab2e-bfa0-63d3-a0c5-c59fb6af57b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.557994] env[63293]: DEBUG oslo_concurrency.lockutils [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.888s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.560555] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.736s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.560834] env[63293]: DEBUG nova.objects.instance [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Lazy-loading 'resources' on Instance uuid 48603812-c3cc-4dae-b8ba-99b9ac5f7969 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.589845] env[63293]: INFO nova.scheduler.client.report [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted allocations for instance c26d6a3a-3643-4b52-870a-4e2ed901d908 [ 876.654914] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327772, 'name': Rename_Task, 'duration_secs': 0.145093} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.655538] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.655847] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d623222f-f450-4710-a979-dc5677392596 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.661529] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for the task: (returnval){ [ 876.661529] env[63293]: value = "task-1327774" [ 876.661529] env[63293]: _type = "Task" [ 876.661529] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.672280] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.692342] env[63293]: DEBUG nova.compute.manager [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.718914] env[63293]: DEBUG nova.virt.hardware [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.719189] env[63293]: DEBUG nova.virt.hardware [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.719422] env[63293]: DEBUG nova.virt.hardware [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.719645] env[63293]: DEBUG nova.virt.hardware [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.719800] env[63293]: DEBUG nova.virt.hardware [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.719952] env[63293]: DEBUG nova.virt.hardware [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.720176] env[63293]: DEBUG nova.virt.hardware [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.720341] env[63293]: DEBUG nova.virt.hardware [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.720524] env[63293]: DEBUG nova.virt.hardware [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.720687] env[63293]: DEBUG nova.virt.hardware [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.720862] env[63293]: DEBUG nova.virt.hardware [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.722211] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5783295-0833-454f-a2c5-a4cff1b4e5f9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.731899] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663978fa-67f2-46cf-aa8d-a1b386b0fbe1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.746690] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.746997] env[63293]: DEBUG nova.compute.manager [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Instance network_info: |[{"id": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "address": "fa:16:3e:12:25:eb", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46d5563f-41", "ovs_interfaceid": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 876.747642] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:25:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aec0089a-ff85-4bef-bad8-c84de39af71a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '46d5563f-41a2-48ab-9d32-3aa394a13da8', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.755846] env[63293]: DEBUG oslo.service.loopingcall [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.756135] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.756410] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ecd577c-0d49-4df7-902e-05687a05de8a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.776261] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.776261] env[63293]: value = "task-1327775" [ 876.776261] env[63293]: _type = "Task" [ 876.776261] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.784372] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327775, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.924542] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a3bb00-f9aa-95bd-3625-cac7d00138fc, 'name': SearchDatastore_Task, 'duration_secs': 0.008224} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.928699] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23d35265-d532-4e7a-9d00-6db8679b5332 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.935223] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 876.935223] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5222d548-0e0d-bbd9-4ab3-eab51efb4b8f" [ 876.935223] env[63293]: _type = "Task" [ 876.935223] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.943518] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5222d548-0e0d-bbd9-4ab3-eab51efb4b8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.952646] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] VM already powered off {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 876.952841] env[63293]: DEBUG nova.compute.manager [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 876.953691] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b41dcb5-e0bb-4295-8b95-73382e549afc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.959258] env[63293]: DEBUG oslo_concurrency.lockutils [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.959500] env[63293]: DEBUG oslo_concurrency.lockutils [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.959692] env[63293]: DEBUG nova.network.neutron [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.011802] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524aab2e-bfa0-63d3-a0c5-c59fb6af57b9, 'name': SearchDatastore_Task, 'duration_secs': 0.008797} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.012135] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.012404] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.012786] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.096270] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "ad585ebb-2072-45df-b645-94c9fa93576b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.096596] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.096818] env[63293]: INFO nova.compute.manager [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Shelving [ 877.101451] env[63293]: DEBUG oslo_concurrency.lockutils [None req-304a4095-7f35-435e-a494-041444a30ea4 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "c26d6a3a-3643-4b52-870a-4e2ed901d908" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.451s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.118941] env[63293]: DEBUG nova.network.neutron [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Successfully updated port: dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.175531] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327774, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.269123] env[63293]: DEBUG nova.objects.instance [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lazy-loading 'flavor' on Instance uuid 831d7d9b-364f-4c29-bda5-e6c38291a973 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.288954] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327775, 'name': CreateVM_Task, 'duration_secs': 0.338793} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.292202] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.293145] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.293278] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.293603] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.294131] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b17a84b-c4c3-40b4-91d2-2b6fc3e6ca14 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.298859] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 877.298859] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f915cb-7c6f-4864-1ac6-27d3b0e4b2d6" [ 877.298859] env[63293]: _type = "Task" [ 877.298859] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.308929] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f915cb-7c6f-4864-1ac6-27d3b0e4b2d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.314261] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459b30ba-be4f-4b26-9694-609d49f6ccce {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.320719] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab7cecb-b0b5-4f63-89fe-1197228f540b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.350407] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d6117e9-352f-4e58-965b-1370b1bd4c1f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.358077] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e048967d-819b-4e02-abbc-ccde25dc7092 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.373737] env[63293]: DEBUG nova.compute.provider_tree [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.432860] env[63293]: DEBUG oslo_concurrency.lockutils [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Acquiring lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.432970] env[63293]: DEBUG oslo_concurrency.lockutils [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.433124] env[63293]: DEBUG oslo_concurrency.lockutils [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Acquiring lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.433316] env[63293]: DEBUG oslo_concurrency.lockutils [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.433490] env[63293]: DEBUG oslo_concurrency.lockutils [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.439867] env[63293]: INFO nova.compute.manager [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Terminating instance [ 877.442040] env[63293]: DEBUG nova.compute.manager [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 877.442321] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0df4ca7f-f4bc-4429-89ad-87db9f724dcc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.447842] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5222d548-0e0d-bbd9-4ab3-eab51efb4b8f, 'name': SearchDatastore_Task, 'duration_secs': 0.009236} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.448467] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.448770] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 828a1165-3829-431c-9edb-3a5d3a8054ec/828a1165-3829-431c-9edb-3a5d3a8054ec.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 877.449064] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.449257] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.449535] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17f7e041-f621-4a6c-8c2b-ea6d19191263 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.452111] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2316940c-f31e-4996-b8d4-e8c30a472849 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.456873] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5b4fcc-fecd-4d40-a2f1-7be560ff252d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.470306] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 877.470306] env[63293]: value = "task-1327776" [ 877.470306] env[63293]: _type = "Task" [ 877.470306] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.474759] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.474945] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.475907] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9389a684-4a66-4eac-ba26-432a2a8a416f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.481024] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327776, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.490289] env[63293]: WARNING nova.virt.vmwareapi.driver [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance b8637881-9fb6-4c5d-848d-7b2d38f8e970 could not be found. [ 877.490491] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 877.490795] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2ea07bc-881c-4924-ba78-bba0be1cf2a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.493321] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 877.493321] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529e0582-ae24-7249-d443-c074c3bee563" [ 877.493321] env[63293]: _type = "Task" [ 877.493321] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.500141] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5944b365-3785-4f36-b9b8-d645512523b5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.515638] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529e0582-ae24-7249-d443-c074c3bee563, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.535689] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b8637881-9fb6-4c5d-848d-7b2d38f8e970 could not be found. [ 877.535974] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 877.536106] env[63293]: INFO nova.compute.manager [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Took 0.09 seconds to destroy the instance on the hypervisor. [ 877.536403] env[63293]: DEBUG oslo.service.loopingcall [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.536604] env[63293]: DEBUG nova.compute.manager [-] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 877.536697] env[63293]: DEBUG nova.network.neutron [-] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 877.607552] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 877.607873] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f670b258-1c80-4d2c-a87f-da62001b1763 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.615652] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 877.615652] env[63293]: value = "task-1327777" [ 877.615652] env[63293]: _type = "Task" [ 877.615652] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.625681] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "refresh_cache-eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.625882] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquired lock "refresh_cache-eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.625992] env[63293]: DEBUG nova.network.neutron [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.632709] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327777, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.673776] env[63293]: DEBUG oslo_vmware.api [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327774, 'name': PowerOnVM_Task, 'duration_secs': 0.534398} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.674116] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.674356] env[63293]: INFO nova.compute.manager [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Took 8.36 seconds to spawn the instance on the hypervisor. [ 877.674545] env[63293]: DEBUG nova.compute.manager [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.675417] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb5022f-b3bc-47fe-81c0-3cbdc781db75 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.686037] env[63293]: DEBUG nova.compute.manager [req-11ba6bfd-d26e-478e-809d-9f0d1bb7e2cc req-fd358b94-2a5f-4c83-a868-088d741ef7e1 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Received event network-vif-plugged-dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.686183] env[63293]: DEBUG oslo_concurrency.lockutils [req-11ba6bfd-d26e-478e-809d-9f0d1bb7e2cc req-fd358b94-2a5f-4c83-a868-088d741ef7e1 service nova] Acquiring lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.686397] env[63293]: DEBUG oslo_concurrency.lockutils [req-11ba6bfd-d26e-478e-809d-9f0d1bb7e2cc req-fd358b94-2a5f-4c83-a868-088d741ef7e1 service nova] Lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.686608] env[63293]: DEBUG oslo_concurrency.lockutils [req-11ba6bfd-d26e-478e-809d-9f0d1bb7e2cc req-fd358b94-2a5f-4c83-a868-088d741ef7e1 service nova] Lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.686793] env[63293]: DEBUG nova.compute.manager [req-11ba6bfd-d26e-478e-809d-9f0d1bb7e2cc req-fd358b94-2a5f-4c83-a868-088d741ef7e1 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] No waiting events found dispatching network-vif-plugged-dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.686965] env[63293]: WARNING nova.compute.manager [req-11ba6bfd-d26e-478e-809d-9f0d1bb7e2cc req-fd358b94-2a5f-4c83-a868-088d741ef7e1 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Received unexpected event network-vif-plugged-dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77 for instance with vm_state building and task_state spawning. [ 877.735489] env[63293]: DEBUG nova.network.neutron [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Successfully updated port: e777ae42-3549-4573-866c-8353417608f1 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.776579] env[63293]: DEBUG oslo_concurrency.lockutils [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquiring lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.776579] env[63293]: DEBUG oslo_concurrency.lockutils [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquired lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.811904] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f915cb-7c6f-4864-1ac6-27d3b0e4b2d6, 'name': SearchDatastore_Task, 'duration_secs': 0.022467} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.812331] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.812696] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.812959] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.877039] env[63293]: DEBUG nova.scheduler.client.report [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.886463] env[63293]: DEBUG nova.network.neutron [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Updating instance_info_cache with network_info: [{"id": "ee3d149b-d66e-439b-a49e-6f2e6c9d9495", "address": "fa:16:3e:30:5d:6f", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee3d149b-d6", "ovs_interfaceid": "ee3d149b-d66e-439b-a49e-6f2e6c9d9495", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.982169] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327776, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.008721] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529e0582-ae24-7249-d443-c074c3bee563, 'name': SearchDatastore_Task, 'duration_secs': 0.009371} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.009988] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db4a0a71-3473-44d4-bd1b-8c8dc3f0ab09 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.020528] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 878.020528] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5262a765-28cb-2170-a468-d723793291d7" [ 878.020528] env[63293]: _type = "Task" [ 878.020528] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.034111] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5262a765-28cb-2170-a468-d723793291d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.101522] env[63293]: DEBUG nova.compute.manager [req-2dc9a700-d409-4802-9a64-0c5a560f1cf9 req-98fbbf7c-11c9-41fd-a591-fea35fada73a service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Received event network-changed-46d5563f-41a2-48ab-9d32-3aa394a13da8 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.101665] env[63293]: DEBUG nova.compute.manager [req-2dc9a700-d409-4802-9a64-0c5a560f1cf9 req-98fbbf7c-11c9-41fd-a591-fea35fada73a service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Refreshing instance network info cache due to event network-changed-46d5563f-41a2-48ab-9d32-3aa394a13da8. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.101889] env[63293]: DEBUG oslo_concurrency.lockutils [req-2dc9a700-d409-4802-9a64-0c5a560f1cf9 req-98fbbf7c-11c9-41fd-a591-fea35fada73a service nova] Acquiring lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.102087] env[63293]: DEBUG oslo_concurrency.lockutils [req-2dc9a700-d409-4802-9a64-0c5a560f1cf9 req-98fbbf7c-11c9-41fd-a591-fea35fada73a service nova] Acquired lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.102262] env[63293]: DEBUG nova.network.neutron [req-2dc9a700-d409-4802-9a64-0c5a560f1cf9 req-98fbbf7c-11c9-41fd-a591-fea35fada73a service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Refreshing network info cache for port 46d5563f-41a2-48ab-9d32-3aa394a13da8 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 878.126510] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327777, 'name': PowerOffVM_Task, 'duration_secs': 0.248246} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.126795] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 878.127706] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22d2a9c-6c52-4ba8-ba44-f771e1e89ebd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.153608] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7026b57-9873-4da9-aab3-76a7e78e582a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.200624] env[63293]: DEBUG nova.network.neutron [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 878.205867] env[63293]: INFO nova.compute.manager [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Took 31.94 seconds to build instance. [ 878.241122] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.241122] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.241122] env[63293]: DEBUG nova.network.neutron [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 878.348318] env[63293]: DEBUG nova.network.neutron [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Updating instance_info_cache with network_info: [{"id": "dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77", "address": "fa:16:3e:91:76:e3", "network": {"id": "02c216f2-c41a-4722-91df-330cd520be3d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.233", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f57886847a1446109c6ab061a248dcf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd9b3d1b-eb", "ovs_interfaceid": "dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.351785] env[63293]: DEBUG nova.network.neutron [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 878.386632] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.826s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.391027] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.350s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.391027] env[63293]: DEBUG nova.objects.instance [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Lazy-loading 'resources' on Instance uuid 4593cabd-1331-4ee1-99f5-46edeac56859 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.394653] env[63293]: DEBUG oslo_concurrency.lockutils [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.410850] env[63293]: INFO nova.scheduler.client.report [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Deleted allocations for instance 48603812-c3cc-4dae-b8ba-99b9ac5f7969 [ 878.483863] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327776, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.67592} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.484168] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 828a1165-3829-431c-9edb-3a5d3a8054ec/828a1165-3829-431c-9edb-3a5d3a8054ec.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 878.484399] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.484676] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9af1449e-218d-47c4-80ba-c07256b443ec {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.494099] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 878.494099] env[63293]: value = "task-1327778" [ 878.494099] env[63293]: _type = "Task" [ 878.494099] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.503351] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327778, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.529212] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5262a765-28cb-2170-a468-d723793291d7, 'name': SearchDatastore_Task, 'duration_secs': 0.056718} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.529481] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.529758] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a/5f01ce51-1f5d-44ad-97f0-2306cbf55b9a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.530081] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.530330] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.530600] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6add54e-ed5d-42df-80f1-753147cfe5a2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.534045] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a12ef985-d79b-4e6d-8ece-7836e461e359 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.542660] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 878.542660] env[63293]: value = "task-1327779" [ 878.542660] env[63293]: _type = "Task" [ 878.542660] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.549228] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.549577] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.553023] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5dc63034-4682-4442-ac11-f15d758fdfb3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.558221] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327779, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.562945] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 878.562945] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524d97fc-f662-ed5b-7178-b914c207546b" [ 878.562945] env[63293]: _type = "Task" [ 878.562945] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.571876] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524d97fc-f662-ed5b-7178-b914c207546b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.665793] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Creating Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 878.665793] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a69e071e-07c7-4833-a063-b6ff02be0b2c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.674026] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 878.674026] env[63293]: value = "task-1327780" [ 878.674026] env[63293]: _type = "Task" [ 878.674026] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.684285] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327780, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.691918] env[63293]: DEBUG nova.network.neutron [-] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.708505] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1bb4629a-2bab-4148-8d3d-81becfde4d18 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.484s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.802687] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 878.803660] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7268478f-22aa-48bd-8c90-154b441c7642 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.808105] env[63293]: WARNING nova.network.neutron [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] 908b99b3-6031-4de0-86a4-09a9cbe355a5 already exists in list: networks containing: ['908b99b3-6031-4de0-86a4-09a9cbe355a5']. ignoring it [ 878.808321] env[63293]: WARNING nova.network.neutron [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] 908b99b3-6031-4de0-86a4-09a9cbe355a5 already exists in list: networks containing: ['908b99b3-6031-4de0-86a4-09a9cbe355a5']. ignoring it [ 878.817689] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 878.818100] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c37c6d1-67d7-4217-b0b7-763bd6e892a6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.851296] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Releasing lock "refresh_cache-eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.851616] env[63293]: DEBUG nova.compute.manager [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Instance network_info: |[{"id": "dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77", "address": "fa:16:3e:91:76:e3", "network": {"id": "02c216f2-c41a-4722-91df-330cd520be3d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.233", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f57886847a1446109c6ab061a248dcf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd9b3d1b-eb", "ovs_interfaceid": "dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.852068] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:76:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.861775] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Creating folder: Project (1cf310a9134147caa91f6aec4ebde76e). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.864368] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a285184-d6ba-4fab-b8fa-4daae523a21e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.876495] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Created folder: Project (1cf310a9134147caa91f6aec4ebde76e) in parent group-v283678. [ 878.876743] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Creating folder: Instances. Parent ref: group-v283778. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.876986] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a337569e-c1a9-4de0-985e-8cc3f13f8825 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.886677] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Created folder: Instances in parent group-v283778. [ 878.887066] env[63293]: DEBUG oslo.service.loopingcall [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.887374] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 878.887614] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63703b47-4b94-4436-afd2-e59868b357a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.917712] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 878.917939] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 878.918145] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleting the datastore file [datastore2] 6cd8b0eb-4bfe-49e0-b392-900f50865ad9 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 878.918667] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d42c203-c32c-470f-9cc2-8bf052ae3f98 tempest-ServerRescueTestJSONUnderV235-1778843507 tempest-ServerRescueTestJSONUnderV235-1778843507-project-member] Lock "48603812-c3cc-4dae-b8ba-99b9ac5f7969" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.181s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.919573] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53107dea-1f2b-442b-a312-5fa2a23f47b9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.923575] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.923575] env[63293]: value = "task-1327784" [ 878.923575] env[63293]: _type = "Task" [ 878.923575] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.929491] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 878.929491] env[63293]: value = "task-1327785" [ 878.929491] env[63293]: _type = "Task" [ 878.929491] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.938614] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327784, 'name': CreateVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.947235] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327785, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.964386] env[63293]: DEBUG nova.network.neutron [req-2dc9a700-d409-4802-9a64-0c5a560f1cf9 req-98fbbf7c-11c9-41fd-a591-fea35fada73a service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updated VIF entry in instance network info cache for port 46d5563f-41a2-48ab-9d32-3aa394a13da8. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 878.964795] env[63293]: DEBUG nova.network.neutron [req-2dc9a700-d409-4802-9a64-0c5a560f1cf9 req-98fbbf7c-11c9-41fd-a591-fea35fada73a service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance_info_cache with network_info: [{"id": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "address": "fa:16:3e:12:25:eb", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46d5563f-41", "ovs_interfaceid": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.011323] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327778, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.162098} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.011323] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.011323] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395ec1ad-e9aa-4d97-8fbb-41124eeb606b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.037588] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 828a1165-3829-431c-9edb-3a5d3a8054ec/828a1165-3829-431c-9edb-3a5d3a8054ec.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.041069] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-076f7250-570b-48fa-a39f-9ca61c5d7a9d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.077620] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 879.077620] env[63293]: value = "task-1327786" [ 879.077620] env[63293]: _type = "Task" [ 879.077620] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.091886] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524d97fc-f662-ed5b-7178-b914c207546b, 'name': SearchDatastore_Task, 'duration_secs': 0.015811} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.091886] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327779, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.095913] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16f9d0ea-088a-419b-b577-c59d2ee940b8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.102491] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327786, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.105879] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 879.105879] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52113598-f7cc-82da-ab93-ccca64a0cddc" [ 879.105879] env[63293]: _type = "Task" [ 879.105879] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.118975] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52113598-f7cc-82da-ab93-ccca64a0cddc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.186798] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327780, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.194671] env[63293]: INFO nova.compute.manager [-] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Took 1.66 seconds to deallocate network for instance. [ 879.291684] env[63293]: DEBUG nova.network.neutron [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updating instance_info_cache with network_info: [{"id": "52a5b959-25ef-48ed-ad04-82f85a131929", "address": "fa:16:3e:56:4e:4a", "network": {"id": "b6024f78-d9d9-4fd8-a71e-9f996ce4a30f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-959509061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76803a169e7d424ab6aefe0d772b7827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52a5b959-25", "ovs_interfaceid": "52a5b959-25ef-48ed-ad04-82f85a131929", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.298171] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae013a0-b009-4ec2-b6d7-2db8282a9f30 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.307689] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ccf4e66-4589-4808-b4ca-c7ec060a7476 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.348532] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d7f36a-9027-437a-b868-97d600baf65f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.358104] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc40e640-225e-44c5-88dd-1121f5afb192 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.375584] env[63293]: DEBUG nova.compute.provider_tree [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.434187] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327784, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.443128] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327785, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.469024] env[63293]: DEBUG oslo_concurrency.lockutils [req-2dc9a700-d409-4802-9a64-0c5a560f1cf9 req-98fbbf7c-11c9-41fd-a591-fea35fada73a service nova] Releasing lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.492680] env[63293]: DEBUG nova.network.neutron [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [{"id": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "address": "fa:16:3e:2b:1c:64", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b2cc4-af", "ovs_interfaceid": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "50288bf0-633c-4e61-a850-e343398936c4", "address": "fa:16:3e:97:22:e5", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50288bf0-63", "ovs_interfaceid": "50288bf0-633c-4e61-a850-e343398936c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e777ae42-3549-4573-866c-8353417608f1", "address": "fa:16:3e:24:f4:eb", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape777ae42-35", "ovs_interfaceid": "e777ae42-3549-4573-866c-8353417608f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.574551] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327779, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.898896} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.574823] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a/5f01ce51-1f5d-44ad-97f0-2306cbf55b9a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.575051] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.575324] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-147d5a6d-f435-4936-974f-058a7f2a4fd9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.586735] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327786, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.588123] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 879.588123] env[63293]: value = "task-1327787" [ 879.588123] env[63293]: _type = "Task" [ 879.588123] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.595946] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327787, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.616582] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52113598-f7cc-82da-ab93-ccca64a0cddc, 'name': SearchDatastore_Task, 'duration_secs': 0.084369} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.616582] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.616859] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 76ebe2ec-332c-4d3a-957e-d0c928dafdc7/76ebe2ec-332c-4d3a-957e-d0c928dafdc7.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 879.617082] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b496f1a8-08c1-4e25-aea9-fb70d3cdd1fc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.624198] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 879.624198] env[63293]: value = "task-1327788" [ 879.624198] env[63293]: _type = "Task" [ 879.624198] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.633527] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.686263] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327780, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.758786] env[63293]: INFO nova.compute.manager [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Took 0.56 seconds to detach 1 volumes for instance. [ 879.761894] env[63293]: DEBUG nova.compute.manager [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Deleting volume: 1f5a039e-f604-47d3-a580-f1754a5b697d {{(pid=63293) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 879.795794] env[63293]: DEBUG oslo_concurrency.lockutils [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Releasing lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.796345] env[63293]: DEBUG nova.compute.manager [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Inject network info {{(pid=63293) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 879.796577] env[63293]: DEBUG nova.compute.manager [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] network_info to inject: |[{"id": "52a5b959-25ef-48ed-ad04-82f85a131929", "address": "fa:16:3e:56:4e:4a", "network": {"id": "b6024f78-d9d9-4fd8-a71e-9f996ce4a30f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-959509061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76803a169e7d424ab6aefe0d772b7827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52a5b959-25", "ovs_interfaceid": "52a5b959-25ef-48ed-ad04-82f85a131929", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 879.803384] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Reconfiguring VM instance to set the machine id {{(pid=63293) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 879.803384] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c10eac38-d450-4582-860f-81e3129e07d7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.825618] env[63293]: DEBUG oslo_vmware.api [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 879.825618] env[63293]: value = "task-1327789" [ 879.825618] env[63293]: _type = "Task" [ 879.825618] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.834628] env[63293]: DEBUG oslo_vmware.api [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327789, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.884799] env[63293]: DEBUG nova.scheduler.client.report [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.888586] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.888757] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.935157] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327784, 'name': CreateVM_Task, 'duration_secs': 0.864527} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.938469] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 879.939702] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.939870] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.940548] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.940909] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ead0ddc1-75b8-476c-ae37-7c285a5b423b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.946751] env[63293]: DEBUG oslo_vmware.api [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327785, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.721696} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.947345] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 879.947564] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 879.947742] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 879.951789] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 879.951789] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ea1bda-073b-9995-8f2b-c6a8d464b34e" [ 879.951789] env[63293]: _type = "Task" [ 879.951789] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.962451] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ea1bda-073b-9995-8f2b-c6a8d464b34e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.975192] env[63293]: INFO nova.scheduler.client.report [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleted allocations for instance 6cd8b0eb-4bfe-49e0-b392-900f50865ad9 [ 879.996348] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.997477] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.997477] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.997780] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796b98b7-db00-4fa2-bff6-ea2d7318927f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.016628] env[63293]: DEBUG nova.virt.hardware [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.016855] env[63293]: DEBUG nova.virt.hardware [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.017685] env[63293]: DEBUG nova.virt.hardware [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.017685] env[63293]: DEBUG nova.virt.hardware [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.017685] env[63293]: DEBUG nova.virt.hardware [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.017685] env[63293]: DEBUG nova.virt.hardware [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.017685] env[63293]: DEBUG nova.virt.hardware [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.018105] env[63293]: DEBUG nova.virt.hardware [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.018105] env[63293]: DEBUG nova.virt.hardware [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.018192] env[63293]: DEBUG nova.virt.hardware [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.018381] env[63293]: DEBUG nova.virt.hardware [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.025938] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Reconfiguring VM to attach interface {{(pid=63293) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 880.027602] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-765593d5-d274-47be-a9e5-91299d353b3b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.047102] env[63293]: DEBUG oslo_vmware.api [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 880.047102] env[63293]: value = "task-1327791" [ 880.047102] env[63293]: _type = "Task" [ 880.047102] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.057944] env[63293]: DEBUG oslo_vmware.api [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327791, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.087783] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327786, 'name': ReconfigVM_Task, 'duration_secs': 0.723244} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.091023] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 828a1165-3829-431c-9edb-3a5d3a8054ec/828a1165-3829-431c-9edb-3a5d3a8054ec.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.091023] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e948368-0082-49d6-a809-2e0420a2cbbd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.101020] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327787, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108155} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.102707] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.103986] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 880.103986] env[63293]: value = "task-1327792" [ 880.103986] env[63293]: _type = "Task" [ 880.103986] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.106556] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c665ffe-7905-48d3-94f3-f4fe5b5b53f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.879813] env[63293]: DEBUG oslo_concurrency.lockutils [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.880569] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.491s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.882662] env[63293]: DEBUG nova.compute.manager [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.887760] env[63293]: DEBUG oslo_concurrency.lockutils [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.888334] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquiring lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.888667] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.889009] env[63293]: INFO nova.compute.manager [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Rebooting instance [ 880.903896] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a/5f01ce51-1f5d-44ad-97f0-2306cbf55b9a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.912580] env[63293]: DEBUG nova.compute.manager [req-56e4ea65-adba-4e02-b0d4-aa9901b150a5 req-c4723c04-3bca-4231-8d96-12ffdc671578 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Received event network-changed-dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.912580] env[63293]: DEBUG nova.compute.manager [req-56e4ea65-adba-4e02-b0d4-aa9901b150a5 req-c4723c04-3bca-4231-8d96-12ffdc671578 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Refreshing instance network info cache due to event network-changed-dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.912840] env[63293]: DEBUG oslo_concurrency.lockutils [req-56e4ea65-adba-4e02-b0d4-aa9901b150a5 req-c4723c04-3bca-4231-8d96-12ffdc671578 service nova] Acquiring lock "refresh_cache-eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.912997] env[63293]: DEBUG oslo_concurrency.lockutils [req-56e4ea65-adba-4e02-b0d4-aa9901b150a5 req-c4723c04-3bca-4231-8d96-12ffdc671578 service nova] Acquired lock "refresh_cache-eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.913604] env[63293]: DEBUG nova.network.neutron [req-56e4ea65-adba-4e02-b0d4-aa9901b150a5 req-c4723c04-3bca-4231-8d96-12ffdc671578 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Refreshing network info cache for port dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.916255] env[63293]: DEBUG nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received event network-vif-plugged-e777ae42-3549-4573-866c-8353417608f1 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.916461] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Acquiring lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.916662] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.916832] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.916997] env[63293]: DEBUG nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] No waiting events found dispatching network-vif-plugged-e777ae42-3549-4573-866c-8353417608f1 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.917811] env[63293]: WARNING nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received unexpected event network-vif-plugged-e777ae42-3549-4573-866c-8353417608f1 for instance with vm_state active and task_state None. [ 880.917811] env[63293]: DEBUG nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received event network-changed-e777ae42-3549-4573-866c-8353417608f1 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.917811] env[63293]: DEBUG nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Refreshing instance network info cache due to event network-changed-e777ae42-3549-4573-866c-8353417608f1. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.917811] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Acquiring lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.918016] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Acquired lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.918074] env[63293]: DEBUG nova.network.neutron [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Refreshing network info cache for port e777ae42-3549-4573-866c-8353417608f1 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.928688] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.261s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.928980] env[63293]: DEBUG nova.objects.instance [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Lazy-loading 'resources' on Instance uuid 93080446-fd45-4281-bf6e-35d1d964f41e {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.936517] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c59826c2-22ac-42ce-b33d-04899100ad50 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.951485] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327792, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.951713] env[63293]: WARNING oslo_vmware.common.loopingcall [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] task run outlasted interval by 0.342202 sec [ 880.961481] env[63293]: INFO nova.scheduler.client.report [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Deleted allocations for instance 4593cabd-1331-4ee1-99f5-46edeac56859 [ 880.990495] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ea1bda-073b-9995-8f2b-c6a8d464b34e, 'name': SearchDatastore_Task, 'duration_secs': 0.053017} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.990495] env[63293]: DEBUG oslo_vmware.api [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327789, 'name': ReconfigVM_Task, 'duration_secs': 0.18718} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.990495] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327780, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.990611] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327788, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644014} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.993183] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquiring lock "refresh_cache-a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.993387] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquired lock "refresh_cache-a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.993578] env[63293]: DEBUG nova.network.neutron [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 880.997944] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.997944] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.997944] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.997944] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.997944] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.997944] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-124b7f8d-5939-42c0-8602-07dd095d7ff2 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Reconfigured VM instance to set the machine id {{(pid=63293) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 881.000135] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 76ebe2ec-332c-4d3a-957e-d0c928dafdc7/76ebe2ec-332c-4d3a-957e-d0c928dafdc7.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.001459] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.001459] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 881.001459] env[63293]: value = "task-1327793" [ 881.001459] env[63293]: _type = "Task" [ 881.001459] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.009593] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b635dd5-66af-4627-8777-eb59d272fd04 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.012280] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8a020c2d-1b74-40da-9bc2-08dbb3b19780 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.015200] env[63293]: DEBUG oslo_vmware.api [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327791, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.027292] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327792, 'name': Rename_Task, 'duration_secs': 0.221084} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.028327] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.028596] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-87e09459-710a-4be5-9fff-bae8ad1d5feb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.036745] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327793, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.039709] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 881.039709] env[63293]: value = "task-1327794" [ 881.039709] env[63293]: _type = "Task" [ 881.039709] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.039937] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.040114] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 881.041880] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44709b27-2154-4c9a-9e39-7e9844b47aae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.051527] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 881.051527] env[63293]: value = "task-1327795" [ 881.051527] env[63293]: _type = "Task" [ 881.051527] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.057785] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 881.057785] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c6f020-2a5c-b4db-3f97-127b7c3b8250" [ 881.057785] env[63293]: _type = "Task" [ 881.057785] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.058784] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327794, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.075059] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327795, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.086929] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c6f020-2a5c-b4db-3f97-127b7c3b8250, 'name': SearchDatastore_Task, 'duration_secs': 0.015987} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.090676] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e74c6a2-808b-42a6-bf1f-f7ebd954360d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.102057] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 881.102057] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5282372e-f8ba-911a-8adc-35f13561c7d9" [ 881.102057] env[63293]: _type = "Task" [ 881.102057] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.111271] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5282372e-f8ba-911a-8adc-35f13561c7d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.279947] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053105cd-c50d-47c2-9525-7b3bce2b542f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.290141] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baef621e-1977-45ce-b87b-d6d0fd46f25d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.320034] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697fdefb-2f8b-4b56-b13b-599240c6f60e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.328240] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00518a38-7322-46d3-87f7-6b6d11f680c1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.342857] env[63293]: DEBUG nova.compute.provider_tree [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.390423] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327780, 'name': CreateSnapshot_Task, 'duration_secs': 2.400732} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.393645] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Created Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 881.394425] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b6d6af-cee0-4d18-af9a-3cf0a52ca9e1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.403871] env[63293]: DEBUG oslo_vmware.api [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327791, 'name': ReconfigVM_Task, 'duration_secs': 1.140798} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.407286] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.407537] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Reconfigured VM to attach interface {{(pid=63293) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 881.432115] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.456415] env[63293]: DEBUG nova.network.neutron [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updated VIF entry in instance network info cache for port e777ae42-3549-4573-866c-8353417608f1. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 881.456918] env[63293]: DEBUG nova.network.neutron [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [{"id": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "address": "fa:16:3e:2b:1c:64", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b2cc4-af", "ovs_interfaceid": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "50288bf0-633c-4e61-a850-e343398936c4", "address": "fa:16:3e:97:22:e5", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50288bf0-63", "ovs_interfaceid": "50288bf0-633c-4e61-a850-e343398936c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e777ae42-3549-4573-866c-8353417608f1", "address": "fa:16:3e:24:f4:eb", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape777ae42-35", "ovs_interfaceid": "e777ae42-3549-4573-866c-8353417608f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.494568] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1a7fcc80-c3e3-439d-a6b7-630e34368dca tempest-ServerMetadataNegativeTestJSON-726088336 tempest-ServerMetadataNegativeTestJSON-726088336-project-member] Lock "4593cabd-1331-4ee1-99f5-46edeac56859" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.389s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.524997] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327793, 'name': ReconfigVM_Task, 'duration_secs': 0.400177} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.525285] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a/5f01ce51-1f5d-44ad-97f0-2306cbf55b9a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.525920] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8f43ffa-6ca7-42c7-ab54-d3dd8f36fb7a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.532993] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 881.532993] env[63293]: value = "task-1327796" [ 881.532993] env[63293]: _type = "Task" [ 881.532993] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.541996] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327796, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.550736] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327794, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102227} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.555544] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.556236] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97013e6e-4568-4dab-a4e4-80ddabbfbdd6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.563586] env[63293]: DEBUG oslo_vmware.api [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327795, 'name': PowerOnVM_Task, 'duration_secs': 0.501945} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.572744] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 881.573048] env[63293]: DEBUG nova.compute.manager [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.581691] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 76ebe2ec-332c-4d3a-957e-d0c928dafdc7/76ebe2ec-332c-4d3a-957e-d0c928dafdc7.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.582509] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d872bfe0-eb87-440e-a254-a226a9a9e60d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.585500] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40b1dd5e-adb0-400a-8851-c2682ce59ae4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.612023] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 881.612023] env[63293]: value = "task-1327797" [ 881.612023] env[63293]: _type = "Task" [ 881.612023] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.619471] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5282372e-f8ba-911a-8adc-35f13561c7d9, 'name': SearchDatastore_Task, 'duration_secs': 0.015887} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.622281] env[63293]: DEBUG nova.network.neutron [req-56e4ea65-adba-4e02-b0d4-aa9901b150a5 req-c4723c04-3bca-4231-8d96-12ffdc671578 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Updated VIF entry in instance network info cache for port dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 881.622697] env[63293]: DEBUG nova.network.neutron [req-56e4ea65-adba-4e02-b0d4-aa9901b150a5 req-c4723c04-3bca-4231-8d96-12ffdc671578 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Updating instance_info_cache with network_info: [{"id": "dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77", "address": "fa:16:3e:91:76:e3", "network": {"id": "02c216f2-c41a-4722-91df-330cd520be3d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.233", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f57886847a1446109c6ab061a248dcf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd9b3d1b-eb", "ovs_interfaceid": "dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.624353] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.624720] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] eec9bb88-c5c4-4d8a-8a30-84b8d89e153d/eec9bb88-c5c4-4d8a-8a30-84b8d89e153d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 881.625262] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a0200719-5984-4c0a-a2d9-9bfc4e657829 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.633019] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327797, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.635191] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 881.635191] env[63293]: value = "task-1327798" [ 881.635191] env[63293]: _type = "Task" [ 881.635191] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.644515] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327798, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.777978] env[63293]: DEBUG nova.network.neutron [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Updating instance_info_cache with network_info: [{"id": "fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2", "address": "fa:16:3e:5f:0f:cc", "network": {"id": "9d41bc05-b8ef-407d-b05d-cc50a54820a5", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-447925489-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81abc1c172e44e68bee6e6fb2b74919a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfddcef04-2e", "ovs_interfaceid": "fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.840362] env[63293]: DEBUG nova.objects.instance [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lazy-loading 'flavor' on Instance uuid 831d7d9b-364f-4c29-bda5-e6c38291a973 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.873646] env[63293]: ERROR nova.scheduler.client.report [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] [req-3c7f3cd6-e29f-4a87-9cf0-fd60ae1d80e5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 619725c3-6194-4724-94b1-1dd499be55d0. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3c7f3cd6-e29f-4a87-9cf0-fd60ae1d80e5"}]} [ 881.904406] env[63293]: DEBUG nova.scheduler.client.report [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Refreshing inventories for resource provider 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 881.916018] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7a165bc3-67e2-4b3f-a1c4-03eb77dbad19 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-e777ae42-3549-4573-866c-8353417608f1" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.067s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.921473] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Creating linked-clone VM from snapshot {{(pid=63293) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 881.922162] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6208e733-ae4e-41e2-b7fc-b7de8d8238f7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.926895] env[63293]: DEBUG nova.scheduler.client.report [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Updating ProviderTree inventory for provider 619725c3-6194-4724-94b1-1dd499be55d0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 881.927305] env[63293]: DEBUG nova.compute.provider_tree [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.937247] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 881.937247] env[63293]: value = "task-1327799" [ 881.937247] env[63293]: _type = "Task" [ 881.937247] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.945162] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327799, 'name': CloneVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.946143] env[63293]: DEBUG nova.scheduler.client.report [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Refreshing aggregate associations for resource provider 619725c3-6194-4724-94b1-1dd499be55d0, aggregates: None {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 881.961437] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Releasing lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.961437] env[63293]: DEBUG nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Received event network-changed-52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.961437] env[63293]: DEBUG nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Refreshing instance network info cache due to event network-changed-52a5b959-25ef-48ed-ad04-82f85a131929. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 881.961437] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Acquiring lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.961437] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Acquired lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.961437] env[63293]: DEBUG nova.network.neutron [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Refreshing network info cache for port 52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.968159] env[63293]: DEBUG nova.scheduler.client.report [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Refreshing trait associations for resource provider 619725c3-6194-4724-94b1-1dd499be55d0, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 882.055452] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327796, 'name': Rename_Task, 'duration_secs': 0.202991} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.055688] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 882.055958] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c7e9ef2-6520-4601-a1d3-b0af0e232f9b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.067204] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 882.067204] env[63293]: value = "task-1327800" [ 882.067204] env[63293]: _type = "Task" [ 882.067204] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.078449] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327800, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.129208] env[63293]: DEBUG oslo_concurrency.lockutils [req-56e4ea65-adba-4e02-b0d4-aa9901b150a5 req-c4723c04-3bca-4231-8d96-12ffdc671578 service nova] Releasing lock "refresh_cache-eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.129568] env[63293]: DEBUG nova.compute.manager [req-56e4ea65-adba-4e02-b0d4-aa9901b150a5 req-c4723c04-3bca-4231-8d96-12ffdc671578 service nova] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Received event network-vif-deleted-aa552093-dd1a-4dd1-aadf-9c92b766b4ea {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.133383] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.137556] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327797, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.146488] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327798, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.272759] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a7b222-4186-4bbb-973c-3b3cc859a1c7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.280748] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Releasing lock "refresh_cache-a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.283415] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d199a8-9704-4a66-866a-b245644b76c2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.287409] env[63293]: DEBUG nova.compute.manager [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.289163] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d7456b-e6ff-440a-bc7e-b63b6319bbe0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.326010] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a0b34c-30d6-4fb4-8b88-0e90aba41807 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.335266] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed50888-1d42-46bf-9994-88b1a2a64b38 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.353718] env[63293]: DEBUG nova.compute.provider_tree [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 882.355301] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquiring lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.445471] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327799, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.583022] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327800, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.628979] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327797, 'name': ReconfigVM_Task, 'duration_secs': 0.818692} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.630533] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 76ebe2ec-332c-4d3a-957e-d0c928dafdc7/76ebe2ec-332c-4d3a-957e-d0c928dafdc7.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.633258] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-891c5349-cc29-4310-a810-f3697bd990da {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.640180] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 882.640180] env[63293]: value = "task-1327801" [ 882.640180] env[63293]: _type = "Task" [ 882.640180] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.661287] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327798, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.749188} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.665107] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] eec9bb88-c5c4-4d8a-8a30-84b8d89e153d/eec9bb88-c5c4-4d8a-8a30-84b8d89e153d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 882.665911] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 882.666323] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327801, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.666667] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-869e45fd-a7db-4e61-8b9c-d0e1772036c2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.673725] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 882.673725] env[63293]: value = "task-1327802" [ 882.673725] env[63293]: _type = "Task" [ 882.673725] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.682829] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327802, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.778089] env[63293]: DEBUG nova.network.neutron [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updated VIF entry in instance network info cache for port 52a5b959-25ef-48ed-ad04-82f85a131929. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.778518] env[63293]: DEBUG nova.network.neutron [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updating instance_info_cache with network_info: [{"id": "52a5b959-25ef-48ed-ad04-82f85a131929", "address": "fa:16:3e:56:4e:4a", "network": {"id": "b6024f78-d9d9-4fd8-a71e-9f996ce4a30f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-959509061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76803a169e7d424ab6aefe0d772b7827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52a5b959-25", "ovs_interfaceid": "52a5b959-25ef-48ed-ad04-82f85a131929", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.894244] env[63293]: DEBUG nova.scheduler.client.report [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Updated inventory for provider 619725c3-6194-4724-94b1-1dd499be55d0 with generation 90 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 882.894244] env[63293]: DEBUG nova.compute.provider_tree [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Updating resource provider 619725c3-6194-4724-94b1-1dd499be55d0 generation from 90 to 91 during operation: update_inventory {{(pid=63293) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 882.894244] env[63293]: DEBUG nova.compute.provider_tree [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 882.945787] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327799, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.083367] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327800, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.160760] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327801, 'name': Rename_Task, 'duration_secs': 0.277543} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.160760] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.160760] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb04c0c4-e5dd-4a03-ba4e-7aead3bc6d0f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.167469] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 883.167469] env[63293]: value = "task-1327803" [ 883.167469] env[63293]: _type = "Task" [ 883.167469] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.175158] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327803, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.183349] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327802, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071501} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.183910] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 883.187019] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09faa853-02c9-48fc-aa95-974429be2e79 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.213657] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] eec9bb88-c5c4-4d8a-8a30-84b8d89e153d/eec9bb88-c5c4-4d8a-8a30-84b8d89e153d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.213657] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20665b1e-3c46-40af-9fdd-23b54c9e7731 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.237765] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 883.237765] env[63293]: value = "task-1327804" [ 883.237765] env[63293]: _type = "Task" [ 883.237765] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.246978] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327804, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.281863] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Releasing lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.282185] env[63293]: DEBUG nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Received event network-vif-unplugged-ee3d149b-d66e-439b-a49e-6f2e6c9d9495 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.282463] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Acquiring lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.282590] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.282757] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.282941] env[63293]: DEBUG nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] No waiting events found dispatching network-vif-unplugged-ee3d149b-d66e-439b-a49e-6f2e6c9d9495 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 883.283131] env[63293]: WARNING nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Received unexpected event network-vif-unplugged-ee3d149b-d66e-439b-a49e-6f2e6c9d9495 for instance with vm_state shelved_offloaded and task_state None. [ 883.283302] env[63293]: DEBUG nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Received event network-changed-ee3d149b-d66e-439b-a49e-6f2e6c9d9495 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.283549] env[63293]: DEBUG nova.compute.manager [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Refreshing instance network info cache due to event network-changed-ee3d149b-d66e-439b-a49e-6f2e6c9d9495. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 883.283763] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Acquiring lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.283906] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Acquired lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.284087] env[63293]: DEBUG nova.network.neutron [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Refreshing network info cache for port ee3d149b-d66e-439b-a49e-6f2e6c9d9495 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 883.285530] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquired lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.315102] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.335516] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399f7dd2-57fa-405a-a2a5-01b395ed1462 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.344123] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Doing hard reboot of VM {{(pid=63293) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 883.344411] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-d61ef286-d027-4f04-aa47-f910fbcbcc8a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.352843] env[63293]: DEBUG oslo_vmware.api [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for the task: (returnval){ [ 883.352843] env[63293]: value = "task-1327805" [ 883.352843] env[63293]: _type = "Task" [ 883.352843] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.361979] env[63293]: DEBUG oslo_vmware.api [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327805, 'name': ResetVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.403488] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.471s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.403488] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 23.225s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.403488] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.403488] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63293) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 883.403652] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.228s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.404195] env[63293]: DEBUG nova.objects.instance [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lazy-loading 'resources' on Instance uuid f427933b-dd2c-49a9-b401-a647635a1eab {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.405871] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b5e85c-7790-4aa1-ac65-f362c7feba33 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.417137] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f2cf84-2fb9-4560-bbf8-b4dda4617abe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.433809] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d486f66a-b11e-48cd-9351-b47f70712f07 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.440033] env[63293]: INFO nova.scheduler.client.report [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Deleted allocations for instance 93080446-fd45-4281-bf6e-35d1d964f41e [ 883.451836] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a633112-92a0-40a2-b05a-9e168c53368e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.462125] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327799, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.498019] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180618MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=63293) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 883.498019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.529196] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquiring lock "828a1165-3829-431c-9edb-3a5d3a8054ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.530050] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "828a1165-3829-431c-9edb-3a5d3a8054ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.530050] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquiring lock "828a1165-3829-431c-9edb-3a5d3a8054ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.530050] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "828a1165-3829-431c-9edb-3a5d3a8054ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.530244] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "828a1165-3829-431c-9edb-3a5d3a8054ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.532590] env[63293]: INFO nova.compute.manager [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Terminating instance [ 883.536021] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquiring lock "refresh_cache-828a1165-3829-431c-9edb-3a5d3a8054ec" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.536645] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquired lock "refresh_cache-828a1165-3829-431c-9edb-3a5d3a8054ec" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.536645] env[63293]: DEBUG nova.network.neutron [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.582353] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327800, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.676271] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327803, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.748827] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327804, 'name': ReconfigVM_Task, 'duration_secs': 0.486689} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.748827] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Reconfigured VM instance instance-0000004f to attach disk [datastore2] eec9bb88-c5c4-4d8a-8a30-84b8d89e153d/eec9bb88-c5c4-4d8a-8a30-84b8d89e153d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.749686] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0fc24f5f-fec3-43c0-bc83-09ed2b398e2e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.756237] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 883.756237] env[63293]: value = "task-1327806" [ 883.756237] env[63293]: _type = "Task" [ 883.756237] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.764753] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327806, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.866124] env[63293]: DEBUG oslo_vmware.api [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327805, 'name': ResetVM_Task, 'duration_secs': 0.112272} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.866479] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Did hard reboot of VM {{(pid=63293) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 883.866731] env[63293]: DEBUG nova.compute.manager [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.868308] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ffbe3f-7e72-4b01-b128-1fc750c20646 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.946264] env[63293]: DEBUG nova.network.neutron [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.959595] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327799, 'name': CloneVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.961570] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3681be7f-d9a5-482c-81cb-fa2364806b95 tempest-ServerAddressesNegativeTestJSON-1484500941 tempest-ServerAddressesNegativeTestJSON-1484500941-project-member] Lock "93080446-fd45-4281-bf6e-35d1d964f41e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.354s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.000751] env[63293]: DEBUG nova.compute.manager [req-f7e0df7e-b83b-4902-93c1-6ab0fbbac9aa req-d38065b3-90d2-48d6-b1e1-66cfc8ecbf2f service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Received event network-changed-52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.000909] env[63293]: DEBUG nova.compute.manager [req-f7e0df7e-b83b-4902-93c1-6ab0fbbac9aa req-d38065b3-90d2-48d6-b1e1-66cfc8ecbf2f service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Refreshing instance network info cache due to event network-changed-52a5b959-25ef-48ed-ad04-82f85a131929. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 884.001137] env[63293]: DEBUG oslo_concurrency.lockutils [req-f7e0df7e-b83b-4902-93c1-6ab0fbbac9aa req-d38065b3-90d2-48d6-b1e1-66cfc8ecbf2f service nova] Acquiring lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.058049] env[63293]: DEBUG nova.network.neutron [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.082728] env[63293]: DEBUG oslo_vmware.api [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327800, 'name': PowerOnVM_Task, 'duration_secs': 1.54317} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.082728] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.082891] env[63293]: INFO nova.compute.manager [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Took 12.33 seconds to spawn the instance on the hypervisor. [ 884.083218] env[63293]: DEBUG nova.compute.manager [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.086543] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafd6d3a-4e26-45fc-8842-c5cec4d260b0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.154431] env[63293]: DEBUG nova.network.neutron [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.160517] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b58c5b-1778-4a9b-89b8-30a521eaa16d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.176074] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993a9148-a4ca-46c7-84d9-15440cd40273 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.189831] env[63293]: DEBUG oslo_vmware.api [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327803, 'name': PowerOnVM_Task, 'duration_secs': 0.693762} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.229509] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.229780] env[63293]: INFO nova.compute.manager [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Took 9.98 seconds to spawn the instance on the hypervisor. [ 884.229966] env[63293]: DEBUG nova.compute.manager [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.231029] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb50c965-d9ae-4842-af4f-25cb2c255ad0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.234244] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6f09f4-d9a1-4424-a2e2-eda4f28abe54 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.244120] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd5b8dc-c8e3-4714-840b-02771497c78f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.263617] env[63293]: DEBUG nova.compute.provider_tree [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.273954] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327806, 'name': Rename_Task, 'duration_secs': 0.153281} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.274888] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 884.275186] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d895d0d7-c6fd-40c7-a57b-58f2949ae594 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.284541] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 884.284541] env[63293]: value = "task-1327807" [ 884.284541] env[63293]: _type = "Task" [ 884.284541] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.294041] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327807, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.380958] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d21b78cf-6345-44fc-b70d-91e1bd261e45 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.492s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.457874] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327799, 'name': CloneVM_Task, 'duration_secs': 2.073277} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.457874] env[63293]: INFO nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Created linked-clone VM from snapshot [ 884.458134] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956816c0-2867-47ed-b74f-1f957f8f68ec {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.468052] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Uploading image dcac372d-af95-4395-b80b-840bcd2b2ed4 {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 884.479978] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-50288bf0-633c-4e61-a850-e343398936c4" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.481650] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-50288bf0-633c-4e61-a850-e343398936c4" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.506380] env[63293]: DEBUG oslo_vmware.rw_handles [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 884.506380] env[63293]: value = "vm-283782" [ 884.506380] env[63293]: _type = "VirtualMachine" [ 884.506380] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 884.506380] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b9592a76-31ac-499b-924f-57d839f51e97 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.527485] env[63293]: DEBUG oslo_vmware.rw_handles [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lease: (returnval){ [ 884.527485] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524bbf00-bcc2-f926-454c-4720adf00565" [ 884.527485] env[63293]: _type = "HttpNfcLease" [ 884.527485] env[63293]: } obtained for exporting VM: (result){ [ 884.527485] env[63293]: value = "vm-283782" [ 884.527485] env[63293]: _type = "VirtualMachine" [ 884.527485] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 884.527485] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the lease: (returnval){ [ 884.527485] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524bbf00-bcc2-f926-454c-4720adf00565" [ 884.527485] env[63293]: _type = "HttpNfcLease" [ 884.527485] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 884.540202] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 884.540202] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524bbf00-bcc2-f926-454c-4720adf00565" [ 884.540202] env[63293]: _type = "HttpNfcLease" [ 884.540202] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 884.610538] env[63293]: INFO nova.compute.manager [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Took 36.12 seconds to build instance. [ 884.633313] env[63293]: DEBUG nova.network.neutron [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Updated VIF entry in instance network info cache for port ee3d149b-d66e-439b-a49e-6f2e6c9d9495. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 884.633744] env[63293]: DEBUG nova.network.neutron [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Updating instance_info_cache with network_info: [{"id": "ee3d149b-d66e-439b-a49e-6f2e6c9d9495", "address": "fa:16:3e:30:5d:6f", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": null, "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapee3d149b-d6", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.657434] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Releasing lock "refresh_cache-828a1165-3829-431c-9edb-3a5d3a8054ec" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.657774] env[63293]: DEBUG nova.compute.manager [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 884.657975] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 884.659624] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a7cd28-f3b9-45ca-ab75-6a30e43df985 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.671337] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 884.671576] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d64218b8-ce60-4569-b0eb-20c037a11cdf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.680419] env[63293]: DEBUG oslo_vmware.api [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 884.680419] env[63293]: value = "task-1327809" [ 884.680419] env[63293]: _type = "Task" [ 884.680419] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.692229] env[63293]: DEBUG oslo_vmware.api [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327809, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.761025] env[63293]: INFO nova.compute.manager [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Took 34.18 seconds to build instance. [ 884.772359] env[63293]: DEBUG nova.scheduler.client.report [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.798096] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327807, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.912912] env[63293]: DEBUG nova.network.neutron [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updating instance_info_cache with network_info: [{"id": "52a5b959-25ef-48ed-ad04-82f85a131929", "address": "fa:16:3e:56:4e:4a", "network": {"id": "b6024f78-d9d9-4fd8-a71e-9f996ce4a30f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-959509061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76803a169e7d424ab6aefe0d772b7827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52a5b959-25", "ovs_interfaceid": "52a5b959-25ef-48ed-ad04-82f85a131929", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.989150] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.989150] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.989150] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0716a9-21a7-47b4-b097-7841b5869f48 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.018698] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1108df07-a455-49b5-9eec-8ae59753f2e5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.021538] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquiring lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.021765] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.021965] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquiring lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.022320] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.022566] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.024953] env[63293]: INFO nova.compute.manager [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Terminating instance [ 885.027626] env[63293]: DEBUG nova.compute.manager [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 885.027626] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 885.029108] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac4823f5-1a4d-46e0-a4b3-8e25654eaf25 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.063109] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Reconfiguring VM to detach interface {{(pid=63293) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 885.067556] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1766cbab-3a4a-42a9-a347-07d6059b3e8f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.083948] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 885.084966] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e64736d6-cf7a-4e56-8be0-6c71ef142288 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.088415] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 885.088415] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524bbf00-bcc2-f926-454c-4720adf00565" [ 885.088415] env[63293]: _type = "HttpNfcLease" [ 885.088415] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 885.089268] env[63293]: DEBUG oslo_vmware.rw_handles [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 885.089268] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524bbf00-bcc2-f926-454c-4720adf00565" [ 885.089268] env[63293]: _type = "HttpNfcLease" [ 885.089268] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 885.090350] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec0fdda-19d5-45c6-8e7a-250cc1f17537 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.093715] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 885.093715] env[63293]: value = "task-1327810" [ 885.093715] env[63293]: _type = "Task" [ 885.093715] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.098163] env[63293]: DEBUG oslo_vmware.api [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for the task: (returnval){ [ 885.098163] env[63293]: value = "task-1327811" [ 885.098163] env[63293]: _type = "Task" [ 885.098163] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.102736] env[63293]: DEBUG oslo_vmware.rw_handles [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf7a9c-b9bd-4380-cafb-7b95f04557be/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 885.103139] env[63293]: DEBUG oslo_vmware.rw_handles [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf7a9c-b9bd-4380-cafb-7b95f04557be/disk-0.vmdk for reading. {{(pid=63293) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 885.112671] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.113164] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e4a01985-ce35-46d6-8588-e8244ab1f7d4 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.580s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.173491] env[63293]: DEBUG oslo_concurrency.lockutils [req-536bcf77-1f04-4866-ab2e-7aecc76edbb9 req-7bc5ded9-5c2b-4aaa-a448-772545f2ad28 service nova] Releasing lock "refresh_cache-6cd8b0eb-4bfe-49e0-b392-900f50865ad9" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.179044] env[63293]: DEBUG oslo_vmware.api [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327811, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.191085] env[63293]: DEBUG oslo_vmware.api [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327809, 'name': PowerOffVM_Task, 'duration_secs': 0.259247} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.191262] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 885.191332] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 885.191607] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f004300-585e-4e07-85d6-ff2ee849b318 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.218309] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f66b8846-2cd4-4838-81c9-d7a5e7af4380 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.223746] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 885.223956] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 885.224202] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Deleting the datastore file [datastore2] 828a1165-3829-431c-9edb-3a5d3a8054ec {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.224502] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a08d52fe-ce6f-4df0-a261-3640a562f12a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.233277] env[63293]: DEBUG oslo_vmware.api [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for the task: (returnval){ [ 885.233277] env[63293]: value = "task-1327813" [ 885.233277] env[63293]: _type = "Task" [ 885.233277] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.244201] env[63293]: DEBUG oslo_vmware.api [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327813, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.272965] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be52b44f-348b-4cf2-afc2-0419cec0675a tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.124s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.278837] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.875s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.281367] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.187s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.281709] env[63293]: DEBUG nova.objects.instance [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lazy-loading 'resources' on Instance uuid 8ea7f3c3-8d73-4fb2-a848-4be4649e5660 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.295527] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327807, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.311179] env[63293]: INFO nova.scheduler.client.report [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Deleted allocations for instance f427933b-dd2c-49a9-b401-a647635a1eab [ 885.419177] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Releasing lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.419734] env[63293]: DEBUG nova.compute.manager [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Inject network info {{(pid=63293) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 885.420443] env[63293]: DEBUG nova.compute.manager [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] network_info to inject: |[{"id": "52a5b959-25ef-48ed-ad04-82f85a131929", "address": "fa:16:3e:56:4e:4a", "network": {"id": "b6024f78-d9d9-4fd8-a71e-9f996ce4a30f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-959509061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76803a169e7d424ab6aefe0d772b7827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52a5b959-25", "ovs_interfaceid": "52a5b959-25ef-48ed-ad04-82f85a131929", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 885.426045] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Reconfiguring VM instance to set the machine id {{(pid=63293) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 885.426799] env[63293]: DEBUG oslo_concurrency.lockutils [req-f7e0df7e-b83b-4902-93c1-6ab0fbbac9aa req-d38065b3-90d2-48d6-b1e1-66cfc8ecbf2f service nova] Acquired lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.427069] env[63293]: DEBUG nova.network.neutron [req-f7e0df7e-b83b-4902-93c1-6ab0fbbac9aa req-d38065b3-90d2-48d6-b1e1-66cfc8ecbf2f service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Refreshing network info cache for port 52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.428414] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-763ecad4-ea0a-4a84-ba50-01a32f3e7f3d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.454563] env[63293]: DEBUG oslo_vmware.api [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 885.454563] env[63293]: value = "task-1327814" [ 885.454563] env[63293]: _type = "Task" [ 885.454563] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.467025] env[63293]: DEBUG oslo_vmware.api [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327814, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.615029] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.624165] env[63293]: DEBUG oslo_vmware.api [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327811, 'name': PowerOffVM_Task, 'duration_secs': 0.227754} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.628457] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 885.629070] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 885.629628] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80a87b5a-088f-41e9-bf7a-ba849e192401 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.719022] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 885.719022] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 885.719022] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Deleting the datastore file [datastore2] a15fc833-d9e8-4340-ab4f-c4a0b2d0e703 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.719022] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf0e13f9-956a-44d9-b026-5c7b67371cc7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.727513] env[63293]: DEBUG oslo_vmware.api [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for the task: (returnval){ [ 885.727513] env[63293]: value = "task-1327816" [ 885.727513] env[63293]: _type = "Task" [ 885.727513] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.751582] env[63293]: DEBUG oslo_vmware.api [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Task: {'id': task-1327813, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172843} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.751582] env[63293]: DEBUG oslo_vmware.api [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327816, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.757297] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.757297] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 885.757297] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 885.757297] env[63293]: INFO nova.compute.manager [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Took 1.10 seconds to destroy the instance on the hypervisor. [ 885.757297] env[63293]: DEBUG oslo.service.loopingcall [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.757297] env[63293]: DEBUG nova.compute.manager [-] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 885.757297] env[63293]: DEBUG nova.network.neutron [-] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 885.761194] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquiring lock "831d7d9b-364f-4c29-bda5-e6c38291a973" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.762390] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lock "831d7d9b-364f-4c29-bda5-e6c38291a973" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.762905] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquiring lock "831d7d9b-364f-4c29-bda5-e6c38291a973-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.763633] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lock "831d7d9b-364f-4c29-bda5-e6c38291a973-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.763633] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lock "831d7d9b-364f-4c29-bda5-e6c38291a973-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.769392] env[63293]: INFO nova.compute.manager [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Terminating instance [ 885.776353] env[63293]: DEBUG nova.compute.manager [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 885.776353] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 885.776353] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ece1c0-f15b-433f-9add-ca7585b1a6f7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.784028] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 885.788557] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f676f51-86b6-4bcf-b4c5-814f35577dea {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.807023] env[63293]: DEBUG oslo_vmware.api [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327807, 'name': PowerOnVM_Task, 'duration_secs': 1.343604} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.807023] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.807563] env[63293]: INFO nova.compute.manager [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Took 9.11 seconds to spawn the instance on the hypervisor. [ 885.807924] env[63293]: DEBUG nova.compute.manager [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.808782] env[63293]: DEBUG oslo_vmware.api [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 885.808782] env[63293]: value = "task-1327817" [ 885.808782] env[63293]: _type = "Task" [ 885.808782] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.809707] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7874bc71-6d06-4a3c-9c35-5c349c289b69 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.816464] env[63293]: DEBUG nova.network.neutron [-] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 885.833427] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ac170c61-cd24-4855-96c9-e311537caf44 tempest-ServersTestMultiNic-892002794 tempest-ServersTestMultiNic-892002794-project-member] Lock "f427933b-dd2c-49a9-b401-a647635a1eab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.876s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.847330] env[63293]: DEBUG oslo_vmware.api [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327817, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.975960] env[63293]: DEBUG oslo_vmware.api [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327814, 'name': ReconfigVM_Task, 'duration_secs': 0.195258} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.981161] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3f6bf513-0d58-4987-a10b-73f1b81fcfe0 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Reconfigured VM instance to set the machine id {{(pid=63293) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 886.109273] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.118130] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d810b0b4-aa2c-46ac-a6ad-1896a5fd5b2d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.128073] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c9cbf51-4fb8-4801-8b3a-5f62055a1339 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.161428] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c493d20-a7bf-4d62-8537-737c16bd93bc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.171921] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebd4cdc-1bdc-4508-8352-a64a20c6deb2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.190619] env[63293]: DEBUG nova.compute.provider_tree [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.240042] env[63293]: DEBUG oslo_vmware.api [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Task: {'id': task-1327816, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247158} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.240893] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.240893] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 886.241176] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 886.241330] env[63293]: INFO nova.compute.manager [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Took 1.21 seconds to destroy the instance on the hypervisor. [ 886.242065] env[63293]: DEBUG oslo.service.loopingcall [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.242065] env[63293]: DEBUG nova.compute.manager [-] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 886.242294] env[63293]: DEBUG nova.network.neutron [-] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 886.326132] env[63293]: DEBUG oslo_vmware.api [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327817, 'name': PowerOffVM_Task, 'duration_secs': 0.253582} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.327161] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 886.327161] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 886.327161] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14f22e21-de50-42bb-a98d-27c0f3a08850 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.332068] env[63293]: DEBUG nova.network.neutron [-] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.349225] env[63293]: INFO nova.compute.manager [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Took 34.81 seconds to build instance. [ 886.403325] env[63293]: DEBUG nova.network.neutron [req-f7e0df7e-b83b-4902-93c1-6ab0fbbac9aa req-d38065b3-90d2-48d6-b1e1-66cfc8ecbf2f service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updated VIF entry in instance network info cache for port 52a5b959-25ef-48ed-ad04-82f85a131929. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 886.403935] env[63293]: DEBUG nova.network.neutron [req-f7e0df7e-b83b-4902-93c1-6ab0fbbac9aa req-d38065b3-90d2-48d6-b1e1-66cfc8ecbf2f service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updating instance_info_cache with network_info: [{"id": "52a5b959-25ef-48ed-ad04-82f85a131929", "address": "fa:16:3e:56:4e:4a", "network": {"id": "b6024f78-d9d9-4fd8-a71e-9f996ce4a30f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-959509061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76803a169e7d424ab6aefe0d772b7827", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52a5b959-25", "ovs_interfaceid": "52a5b959-25ef-48ed-ad04-82f85a131929", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.435545] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 886.435545] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 886.435545] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Deleting the datastore file [datastore1] 831d7d9b-364f-4c29-bda5-e6c38291a973 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.436953] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36d11023-2201-485e-8a10-3afee8333d11 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.447020] env[63293]: DEBUG oslo_vmware.api [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for the task: (returnval){ [ 886.447020] env[63293]: value = "task-1327819" [ 886.447020] env[63293]: _type = "Task" [ 886.447020] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.460721] env[63293]: DEBUG oslo_vmware.api [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327819, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.612342] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.695049] env[63293]: DEBUG nova.scheduler.client.report [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.776727] env[63293]: DEBUG nova.compute.manager [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.777729] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311aac84-8be2-4a7e-8f2c-e8e955981fa0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.834466] env[63293]: INFO nova.compute.manager [-] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Took 1.08 seconds to deallocate network for instance. [ 886.853817] env[63293]: DEBUG oslo_concurrency.lockutils [None req-0f85ebdc-6ee1-40a8-9482-d17edea983e0 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.610s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.907188] env[63293]: DEBUG oslo_concurrency.lockutils [req-f7e0df7e-b83b-4902-93c1-6ab0fbbac9aa req-d38065b3-90d2-48d6-b1e1-66cfc8ecbf2f service nova] Releasing lock "refresh_cache-831d7d9b-364f-4c29-bda5-e6c38291a973" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.963066] env[63293]: DEBUG oslo_vmware.api [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Task: {'id': task-1327819, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190441} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.963494] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.967189] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 886.967189] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 886.967388] env[63293]: INFO nova.compute.manager [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Took 1.20 seconds to destroy the instance on the hypervisor. [ 886.967669] env[63293]: DEBUG oslo.service.loopingcall [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.968013] env[63293]: DEBUG nova.compute.manager [-] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 886.968421] env[63293]: DEBUG nova.network.neutron [-] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 887.112447] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.172866] env[63293]: DEBUG nova.compute.manager [req-764c2231-d685-4a70-8819-354c50476cc7 req-3fda2091-2ca5-4390-ba53-557c0c3cb024 service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Received event network-vif-deleted-fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.173550] env[63293]: INFO nova.compute.manager [req-764c2231-d685-4a70-8819-354c50476cc7 req-3fda2091-2ca5-4390-ba53-557c0c3cb024 service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Neutron deleted interface fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2; detaching it from the instance and deleting it from the info cache [ 887.174084] env[63293]: DEBUG nova.network.neutron [req-764c2231-d685-4a70-8819-354c50476cc7 req-3fda2091-2ca5-4390-ba53-557c0c3cb024 service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.200917] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.920s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.206352] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.733s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.206645] env[63293]: DEBUG nova.objects.instance [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lazy-loading 'resources' on Instance uuid 9973dd7f-ca75-4db0-8d2b-f21b0311abeb {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.241651] env[63293]: INFO nova.scheduler.client.report [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Deleted allocations for instance 8ea7f3c3-8d73-4fb2-a848-4be4649e5660 [ 887.297865] env[63293]: INFO nova.compute.manager [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] instance snapshotting [ 887.301550] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ded460-2ee4-4145-9e68-c00e83f2983d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.332216] env[63293]: DEBUG nova.network.neutron [-] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.336158] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9fc4b0f-153d-4ce2-bb9e-fb054715d74b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.346666] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.612127] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.677371] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5723b69b-396d-4032-91f6-360417698831 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.693976] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cca84f0-edf6-4b19-8d23-3d5ec68e30c5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.738309] env[63293]: DEBUG nova.compute.manager [req-764c2231-d685-4a70-8819-354c50476cc7 req-3fda2091-2ca5-4390-ba53-557c0c3cb024 service nova] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Detach interface failed, port_id=fddcef04-2ee3-44cf-8b72-6e6e0c3f23c2, reason: Instance a15fc833-d9e8-4340-ab4f-c4a0b2d0e703 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 887.756353] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa029de7-bed5-443a-ab1c-330186c1d82e tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "8ea7f3c3-8d73-4fb2-a848-4be4649e5660" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.528s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.841648] env[63293]: INFO nova.compute.manager [-] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Took 1.60 seconds to deallocate network for instance. [ 887.852453] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Creating Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 887.853050] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-17d15929-7074-49f0-8ed8-dddb748f8c60 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.868160] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 887.868160] env[63293]: value = "task-1327820" [ 887.868160] env[63293]: _type = "Task" [ 887.868160] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.884136] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327820, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.982295] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a42487-cd59-45c7-88f9-df7ffaf63e84 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.996769] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a65a9f1-7639-4a48-9e66-867bc0d21cd3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.039634] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87574d0-7e49-44e0-81cd-cf46e6f62149 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.050625] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d97b04-ba09-4f3c-bf59-c07ddf9f2985 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.070018] env[63293]: DEBUG nova.compute.provider_tree [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 888.108882] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.353180] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.381773] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327820, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.470396] env[63293]: DEBUG nova.network.neutron [-] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.598713] env[63293]: ERROR nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [req-ba4afb5c-020c-446b-974e-b47b0afc1839] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 619725c3-6194-4724-94b1-1dd499be55d0. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ba4afb5c-020c-446b-974e-b47b0afc1839"}]} [ 888.616607] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.636372] env[63293]: DEBUG nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Refreshing inventories for resource provider 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 888.665466] env[63293]: DEBUG nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Updating ProviderTree inventory for provider 619725c3-6194-4724-94b1-1dd499be55d0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 888.665744] env[63293]: DEBUG nova.compute.provider_tree [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 888.697665] env[63293]: DEBUG nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Refreshing aggregate associations for resource provider 619725c3-6194-4724-94b1-1dd499be55d0, aggregates: None {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 888.705410] env[63293]: DEBUG nova.compute.manager [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.705410] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe5dd62-20d2-45d5-a70b-d65925a5a4b5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.737843] env[63293]: DEBUG nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Refreshing trait associations for resource provider 619725c3-6194-4724-94b1-1dd499be55d0, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 888.881414] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327820, 'name': CreateSnapshot_Task, 'duration_secs': 0.735922} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.882013] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Created Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 888.882535] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a57eca-1016-4644-9c0e-9b642153f4df {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.974699] env[63293]: INFO nova.compute.manager [-] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Took 2.01 seconds to deallocate network for instance. [ 888.990331] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e72ca3b-fe66-442e-943e-2d2a9a355082 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.998830] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32360dc-6d5c-4a22-a974-762e09b95d40 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.032321] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2e0f2b-c1bd-470d-bb8b-604b5461a823 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.041489] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5473f15-8543-4eaf-82fd-e4b3115be909 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.057479] env[63293]: DEBUG nova.compute.provider_tree [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 889.067682] env[63293]: DEBUG nova.compute.manager [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Stashing vm_state: active {{(pid=63293) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 889.111898] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.220366] env[63293]: INFO nova.compute.manager [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] instance snapshotting [ 889.220366] env[63293]: DEBUG nova.objects.instance [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'flavor' on Instance uuid 9ddf4f23-1279-4cbd-8212-10f344060445 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.229136] env[63293]: DEBUG nova.compute.manager [req-a86f70cd-ad35-4f91-9594-1c03ffa6f6f8 req-99586f9b-496c-4483-8c1c-3958970f9bab service nova] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Received event network-vif-deleted-52a5b959-25ef-48ed-ad04-82f85a131929 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.260599] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "5b1264b4-2554-4c62-9ff9-4616b03d3609" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.260868] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "5b1264b4-2554-4c62-9ff9-4616b03d3609" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.405682] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Creating linked-clone VM from snapshot {{(pid=63293) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 889.405682] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-99463092-52e1-4d2c-a36f-475e91128cc0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.414228] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 889.414228] env[63293]: value = "task-1327821" [ 889.414228] env[63293]: _type = "Task" [ 889.414228] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.423739] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327821, 'name': CloneVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.484163] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.580815] env[63293]: ERROR nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] [req-8203b609-7673-4687-bd6d-00d20ffdb9f8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 619725c3-6194-4724-94b1-1dd499be55d0. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8203b609-7673-4687-bd6d-00d20ffdb9f8"}]} [ 889.588669] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.601051] env[63293]: DEBUG nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Refreshing inventories for resource provider 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 889.621599] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.623288] env[63293]: DEBUG nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Updating ProviderTree inventory for provider 619725c3-6194-4724-94b1-1dd499be55d0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 889.623645] env[63293]: DEBUG nova.compute.provider_tree [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 889.637621] env[63293]: DEBUG nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Refreshing aggregate associations for resource provider 619725c3-6194-4724-94b1-1dd499be55d0, aggregates: None {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 889.667458] env[63293]: DEBUG nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Refreshing trait associations for resource provider 619725c3-6194-4724-94b1-1dd499be55d0, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 889.726825] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d328a4a8-a29c-4fa9-83de-a266e9a4a05e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.747564] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d186a56b-b779-4000-b9bf-975041adbf2c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.763660] env[63293]: DEBUG nova.compute.manager [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 889.925842] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327821, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.971721] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e41a0cd-237b-4eac-bdfe-ad1adbfae111 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.981853] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50719be2-9e9e-40c6-897f-8f1c05d6f8aa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.022383] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b68f999-b488-4446-a01e-79a68b8723d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.028558] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquiring lock "8625a358-ec48-46b9-8738-c49a2ba58362" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.028786] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Lock "8625a358-ec48-46b9-8738-c49a2ba58362" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.037876] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41d7ca3-17f8-4777-b968-70a3c5b1cec9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.053679] env[63293]: DEBUG nova.compute.provider_tree [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 890.114574] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.270345] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Creating Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 890.271832] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6d016c03-1950-4a86-bc63-2291562d8132 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.286232] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 890.286232] env[63293]: value = "task-1327822" [ 890.286232] env[63293]: _type = "Task" [ 890.286232] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.298305] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.298520] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327822, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.428385] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327821, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.531592] env[63293]: DEBUG nova.compute.manager [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.596424] env[63293]: DEBUG nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Updated inventory for provider 619725c3-6194-4724-94b1-1dd499be55d0 with generation 95 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 890.596748] env[63293]: DEBUG nova.compute.provider_tree [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Updating resource provider 619725c3-6194-4724-94b1-1dd499be55d0 generation from 95 to 96 during operation: update_inventory {{(pid=63293) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 890.597806] env[63293]: DEBUG nova.compute.provider_tree [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 890.618278] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.796022] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327822, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.928282] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327821, 'name': CloneVM_Task} progress is 95%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.055716] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.103470] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.897s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.105795] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.028s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.119335] env[63293]: DEBUG oslo_vmware.api [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327810, 'name': ReconfigVM_Task, 'duration_secs': 5.899629} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.120300] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.121192] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Reconfigured VM to detach interface {{(pid=63293) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 891.131888] env[63293]: INFO nova.scheduler.client.report [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Deleted allocations for instance 9973dd7f-ca75-4db0-8d2b-f21b0311abeb [ 891.300504] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327822, 'name': CreateSnapshot_Task, 'duration_secs': 0.835084} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.303452] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Created Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 891.304490] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8031917-bc99-49c0-bdd8-874c90d9f7e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.356644] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44ba558-3c25-4a08-965c-a1b604e3cd7c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.367577] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4397a70-f1cf-4bcf-aaa0-f36ba84d0938 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.405997] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4e06db-cc1a-4b94-b945-3818e1a60666 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.420309] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc2ba74-a252-4545-be70-cd8383824d1f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.432703] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327821, 'name': CloneVM_Task, 'duration_secs': 1.780796} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.441324] env[63293]: INFO nova.virt.vmwareapi.vmops [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Created linked-clone VM from snapshot [ 891.442206] env[63293]: DEBUG nova.compute.provider_tree [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.443935] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e087ab82-c1ad-4c14-b88b-b87a47cee604 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.448879] env[63293]: DEBUG nova.compute.manager [req-fc2f771a-fe31-44ea-9311-63c6f578b508 req-15af217a-0a09-4186-876b-567fc43c0ce1 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received event network-vif-deleted-50288bf0-633c-4e61-a850-e343398936c4 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.448948] env[63293]: INFO nova.compute.manager [req-fc2f771a-fe31-44ea-9311-63c6f578b508 req-15af217a-0a09-4186-876b-567fc43c0ce1 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Neutron deleted interface 50288bf0-633c-4e61-a850-e343398936c4; detaching it from the instance and deleting it from the info cache [ 891.449308] env[63293]: DEBUG nova.network.neutron [req-fc2f771a-fe31-44ea-9311-63c6f578b508 req-15af217a-0a09-4186-876b-567fc43c0ce1 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [{"id": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "address": "fa:16:3e:2b:1c:64", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b2cc4-af", "ovs_interfaceid": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e777ae42-3549-4573-866c-8353417608f1", "address": "fa:16:3e:24:f4:eb", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape777ae42-35", "ovs_interfaceid": "e777ae42-3549-4573-866c-8353417608f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.458663] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Uploading image 8b176c14-05dc-43c1-a475-57283dc96350 {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 891.481313] env[63293]: DEBUG oslo_vmware.rw_handles [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 891.481313] env[63293]: value = "vm-283784" [ 891.481313] env[63293]: _type = "VirtualMachine" [ 891.481313] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 891.481667] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d4a363de-b368-4fa2-9627-074479916edd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.494384] env[63293]: DEBUG oslo_vmware.rw_handles [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lease: (returnval){ [ 891.494384] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f36b40-8afc-7491-99f8-d40d5d575ce8" [ 891.494384] env[63293]: _type = "HttpNfcLease" [ 891.494384] env[63293]: } obtained for exporting VM: (result){ [ 891.494384] env[63293]: value = "vm-283784" [ 891.494384] env[63293]: _type = "VirtualMachine" [ 891.494384] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 891.494663] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the lease: (returnval){ [ 891.494663] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f36b40-8afc-7491-99f8-d40d5d575ce8" [ 891.494663] env[63293]: _type = "HttpNfcLease" [ 891.494663] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 891.501993] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 891.501993] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f36b40-8afc-7491-99f8-d40d5d575ce8" [ 891.501993] env[63293]: _type = "HttpNfcLease" [ 891.501993] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 891.644298] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fe674e94-7a81-4726-996e-2c1a57f1e326 tempest-MultipleCreateTestJSON-373877369 tempest-MultipleCreateTestJSON-373877369-project-member] Lock "9973dd7f-ca75-4db0-8d2b-f21b0311abeb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.098s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.831342] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Creating linked-clone VM from snapshot {{(pid=63293) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 891.831743] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1cb23682-af17-4aa8-847b-74c6ae8af20b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.841646] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 891.841646] env[63293]: value = "task-1327824" [ 891.841646] env[63293]: _type = "Task" [ 891.841646] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.852265] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327824, 'name': CloneVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.949351] env[63293]: DEBUG nova.scheduler.client.report [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.956036] env[63293]: DEBUG oslo_concurrency.lockutils [req-fc2f771a-fe31-44ea-9311-63c6f578b508 req-15af217a-0a09-4186-876b-567fc43c0ce1 service nova] Acquiring lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.956036] env[63293]: DEBUG oslo_concurrency.lockutils [req-fc2f771a-fe31-44ea-9311-63c6f578b508 req-15af217a-0a09-4186-876b-567fc43c0ce1 service nova] Acquired lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.956036] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64fb1a9-c5c4-49c2-a4bd-4d624a3b0aa3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.978784] env[63293]: DEBUG oslo_concurrency.lockutils [req-fc2f771a-fe31-44ea-9311-63c6f578b508 req-15af217a-0a09-4186-876b-567fc43c0ce1 service nova] Releasing lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.979043] env[63293]: WARNING nova.compute.manager [req-fc2f771a-fe31-44ea-9311-63c6f578b508 req-15af217a-0a09-4186-876b-567fc43c0ce1 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Detach interface failed, port_id=50288bf0-633c-4e61-a850-e343398936c4, reason: No device with interface-id 50288bf0-633c-4e61-a850-e343398936c4 exists on VM: nova.exception.NotFound: No device with interface-id 50288bf0-633c-4e61-a850-e343398936c4 exists on VM [ 892.003752] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 892.003752] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f36b40-8afc-7491-99f8-d40d5d575ce8" [ 892.003752] env[63293]: _type = "HttpNfcLease" [ 892.003752] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 892.004300] env[63293]: DEBUG oslo_vmware.rw_handles [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 892.004300] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f36b40-8afc-7491-99f8-d40d5d575ce8" [ 892.004300] env[63293]: _type = "HttpNfcLease" [ 892.004300] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 892.004850] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9e212a-60e8-4be6-a2da-acbe5ab18b04 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.014600] env[63293]: DEBUG oslo_vmware.rw_handles [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f23633-5d1b-4574-1fda-221cdc3fc9a2/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 892.014849] env[63293]: DEBUG oslo_vmware.rw_handles [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f23633-5d1b-4574-1fda-221cdc3fc9a2/disk-0.vmdk for reading. {{(pid=63293) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 892.086051] env[63293]: DEBUG nova.compute.manager [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received event network-vif-deleted-e777ae42-3549-4573-866c-8353417608f1 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.086298] env[63293]: INFO nova.compute.manager [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Neutron deleted interface e777ae42-3549-4573-866c-8353417608f1; detaching it from the instance and deleting it from the info cache [ 892.086568] env[63293]: DEBUG nova.network.neutron [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [{"id": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "address": "fa:16:3e:2b:1c:64", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b2cc4-af", "ovs_interfaceid": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.153647] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9fc6df83-01e8-43a3-a022-367277d1ceef {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.352814] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327824, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.429119] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.429462] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.429701] env[63293]: DEBUG nova.network.neutron [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.458556] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.353s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.459052] env[63293]: INFO nova.compute.manager [None req-b495424c-78ec-4c53-beb1-db5ec5d00401 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Successfully reverted task state from rebuilding on failure for instance. [ 892.464523] env[63293]: DEBUG oslo_concurrency.lockutils [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.585s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.464804] env[63293]: DEBUG nova.objects.instance [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lazy-loading 'resources' on Instance uuid b8637881-9fb6-4c5d-848d-7b2d38f8e970 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.592293] env[63293]: DEBUG oslo_concurrency.lockutils [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Acquiring lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.592562] env[63293]: DEBUG oslo_concurrency.lockutils [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Acquired lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.593036] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.593907] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36395730-0752-4893-9223-b1026dd6e75d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.636362] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96079b2e-26f6-4816-b869-9cc6e5c35c4a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.666865] env[63293]: DEBUG nova.virt.vmwareapi.vmops [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Reconfiguring VM to detach interface {{(pid=63293) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 892.666865] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9520b850-fd33-4ae5-a382-8d1d9dc54e76 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.686869] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Waiting for the task: (returnval){ [ 892.686869] env[63293]: value = "task-1327825" [ 892.686869] env[63293]: _type = "Task" [ 892.686869] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.697622] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.855510] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327824, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.210457] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.232024] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a1b2eb-61fe-4dfd-aeba-b8c1e10cdbaf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.242657] env[63293]: DEBUG nova.network.neutron [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [{"id": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "address": "fa:16:3e:2b:1c:64", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34b2cc4-af", "ovs_interfaceid": "a34b2cc4-afe1-463c-bb22-470ea54d8789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.245197] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd5866e-e782-4c13-870d-a34cf1d99ea9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.285859] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aae6f8b-04a6-4641-be5e-e97bf3c3e652 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.299629] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02252356-8005-454b-9ae9-33e4df457824 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.320341] env[63293]: DEBUG nova.compute.provider_tree [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.361802] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327824, 'name': CloneVM_Task} progress is 95%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.677191] env[63293]: DEBUG oslo_vmware.rw_handles [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf7a9c-b9bd-4380-cafb-7b95f04557be/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 893.678757] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a2b481a-9b49-4528-8820-67b1f759c709 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.689018] env[63293]: DEBUG oslo_vmware.rw_handles [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf7a9c-b9bd-4380-cafb-7b95f04557be/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 893.689018] env[63293]: ERROR oslo_vmware.rw_handles [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf7a9c-b9bd-4380-cafb-7b95f04557be/disk-0.vmdk due to incomplete transfer. [ 893.689018] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-74c29793-441e-41cf-9afe-a7ee6e7c0626 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.702064] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.703598] env[63293]: DEBUG oslo_vmware.rw_handles [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf7a9c-b9bd-4380-cafb-7b95f04557be/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 893.704124] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Uploaded image dcac372d-af95-4395-b80b-840bcd2b2ed4 to the Glance image server {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 893.706347] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Destroying the VM {{(pid=63293) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 893.706765] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d537cd00-79c9-4f18-ac5b-a97e75551dbc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.718018] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 893.718018] env[63293]: value = "task-1327826" [ 893.718018] env[63293]: _type = "Task" [ 893.718018] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.727846] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327826, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.750938] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.823615] env[63293]: DEBUG nova.scheduler.client.report [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.857790] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327824, 'name': CloneVM_Task, 'duration_secs': 1.754084} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.858171] env[63293]: INFO nova.virt.vmwareapi.vmops [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Created linked-clone VM from snapshot [ 893.859199] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eafbfe7-5e95-4198-aa86-e1d57a1e79ce {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.868511] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Uploading image 43bec244-5978-48cb-96cf-849335c88aa8 {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 893.895030] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 893.895030] env[63293]: value = "vm-283786" [ 893.895030] env[63293]: _type = "VirtualMachine" [ 893.895030] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 893.895374] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1e283584-f535-46ae-82e4-08e8ffce12c1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.905394] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lease: (returnval){ [ 893.905394] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52352556-ac35-c798-55e6-1ad4af1a0074" [ 893.905394] env[63293]: _type = "HttpNfcLease" [ 893.905394] env[63293]: } obtained for exporting VM: (result){ [ 893.905394] env[63293]: value = "vm-283786" [ 893.905394] env[63293]: _type = "VirtualMachine" [ 893.905394] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 893.905733] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the lease: (returnval){ [ 893.905733] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52352556-ac35-c798-55e6-1ad4af1a0074" [ 893.905733] env[63293]: _type = "HttpNfcLease" [ 893.905733] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 893.914104] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 893.914104] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52352556-ac35-c798-55e6-1ad4af1a0074" [ 893.914104] env[63293]: _type = "HttpNfcLease" [ 893.914104] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 894.202618] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.232541] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327826, 'name': Destroy_Task, 'duration_secs': 0.35747} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.232934] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Destroyed the VM [ 894.233317] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Deleting Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 894.233699] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f38ca2f3-dcae-4de7-94b6-6588902e6339 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.242174] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 894.242174] env[63293]: value = "task-1327828" [ 894.242174] env[63293]: _type = "Task" [ 894.242174] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.252835] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327828, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.255900] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9ea100af-b116-444b-aba2-e35203f961af tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-65c2f72d-6bbe-45ac-8efc-401e0664390f-50288bf0-633c-4e61-a850-e343398936c4" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.775s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.329788] env[63293]: DEBUG oslo_concurrency.lockutils [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.332327] env[63293]: DEBUG oslo_concurrency.lockutils [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.445s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.332575] env[63293]: DEBUG nova.objects.instance [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lazy-loading 'resources' on Instance uuid 6cd8b0eb-4bfe-49e0-b392-900f50865ad9 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.415529] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 894.415529] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52352556-ac35-c798-55e6-1ad4af1a0074" [ 894.415529] env[63293]: _type = "HttpNfcLease" [ 894.415529] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 894.415929] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 894.415929] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52352556-ac35-c798-55e6-1ad4af1a0074" [ 894.415929] env[63293]: _type = "HttpNfcLease" [ 894.415929] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 894.416710] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2783e2a4-f53c-4d4d-b1b2-7af2d0073353 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.425244] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b7c635-225c-5bdf-4563-bfbf0e7beeb4/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 894.425579] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b7c635-225c-5bdf-4563-bfbf0e7beeb4/disk-0.vmdk for reading. {{(pid=63293) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 894.536106] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-281f5a4d-1370-464a-ab5c-72eb1f0bd155 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.702767] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.754396] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327828, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.838118] env[63293]: DEBUG nova.objects.instance [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lazy-loading 'numa_topology' on Instance uuid 6cd8b0eb-4bfe-49e0-b392-900f50865ad9 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.850103] env[63293]: DEBUG oslo_concurrency.lockutils [None req-381585ee-6f6e-4e68-99ba-3a8707d8d376 tempest-ServerActionsV293TestJSON-548590961 tempest-ServerActionsV293TestJSON-548590961-project-member] Lock "b8637881-9fb6-4c5d-848d-7b2d38f8e970" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.417s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.202498] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.253957] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327828, 'name': RemoveSnapshot_Task, 'duration_secs': 0.68509} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.254355] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Deleted Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 895.254666] env[63293]: DEBUG nova.compute.manager [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.255521] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b13764-63d7-493f-9df0-4c5034ee6733 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.341989] env[63293]: DEBUG nova.objects.base [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Object Instance<6cd8b0eb-4bfe-49e0-b392-900f50865ad9> lazy-loaded attributes: resources,numa_topology {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 895.558479] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa1db4f-cc59-4390-9bc8-c437652d8c57 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.568826] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb64b3c2-978f-40c6-a202-56f25e492a86 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.605650] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404f2ad8-4002-4fa6-ba6b-83b560b7b175 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.616379] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d610aa9-1df7-40d2-9a20-f349fda460cb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.634769] env[63293]: DEBUG nova.compute.provider_tree [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.703061] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.772063] env[63293]: INFO nova.compute.manager [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Shelve offloading [ 895.775600] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.776165] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c7d50f0-ad8a-4a59-b4d9-b8bdf5fa3d1d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.788224] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 895.788224] env[63293]: value = "task-1327829" [ 895.788224] env[63293]: _type = "Task" [ 895.788224] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.798779] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] VM already powered off {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 895.799094] env[63293]: DEBUG nova.compute.manager [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.800009] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39bff939-4c7d-4bc7-bfde-cf751dc72cc2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.807505] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.807505] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.807505] env[63293]: DEBUG nova.network.neutron [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 896.138711] env[63293]: DEBUG nova.scheduler.client.report [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.204060] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.536045] env[63293]: DEBUG nova.network.neutron [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updating instance_info_cache with network_info: [{"id": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "address": "fa:16:3e:c0:e8:15", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0576b708-1b", "ovs_interfaceid": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.648092] env[63293]: DEBUG oslo_concurrency.lockutils [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.314s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.649800] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.218s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.651611] env[63293]: INFO nova.compute.claims [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.705539] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.039115] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.157018] env[63293]: DEBUG oslo_concurrency.lockutils [None req-738f67fb-08f6-4b95-bf72-31f21db75d29 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 36.958s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.160444] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 13.845s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.160713] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.160904] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.161084] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.163894] env[63293]: INFO nova.compute.manager [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Terminating instance [ 897.166203] env[63293]: DEBUG nova.compute.manager [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 897.166203] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 897.166280] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ed0a1b7-7041-4502-ba34-60ab744d8292 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.181098] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d791ebcb-7d1a-4f83-b4af-90ac24a88eac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.204611] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.216670] env[63293]: WARNING nova.virt.vmwareapi.vmops [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6cd8b0eb-4bfe-49e0-b392-900f50865ad9 could not be found. [ 897.217032] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 897.217108] env[63293]: INFO nova.compute.manager [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 897.217527] env[63293]: DEBUG oslo.service.loopingcall [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.217819] env[63293]: DEBUG nova.compute.manager [-] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 897.217916] env[63293]: DEBUG nova.network.neutron [-] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 897.328875] env[63293]: DEBUG nova.compute.manager [req-60dfad11-17d0-4f63-b2af-bfe1fa213371 req-1de1ab28-cd94-409d-b87b-80ca07a8cdf2 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Received event network-vif-unplugged-0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.329129] env[63293]: DEBUG oslo_concurrency.lockutils [req-60dfad11-17d0-4f63-b2af-bfe1fa213371 req-1de1ab28-cd94-409d-b87b-80ca07a8cdf2 service nova] Acquiring lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.329354] env[63293]: DEBUG oslo_concurrency.lockutils [req-60dfad11-17d0-4f63-b2af-bfe1fa213371 req-1de1ab28-cd94-409d-b87b-80ca07a8cdf2 service nova] Lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.329613] env[63293]: DEBUG oslo_concurrency.lockutils [req-60dfad11-17d0-4f63-b2af-bfe1fa213371 req-1de1ab28-cd94-409d-b87b-80ca07a8cdf2 service nova] Lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.329796] env[63293]: DEBUG nova.compute.manager [req-60dfad11-17d0-4f63-b2af-bfe1fa213371 req-1de1ab28-cd94-409d-b87b-80ca07a8cdf2 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] No waiting events found dispatching network-vif-unplugged-0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 897.329973] env[63293]: WARNING nova.compute.manager [req-60dfad11-17d0-4f63-b2af-bfe1fa213371 req-1de1ab28-cd94-409d-b87b-80ca07a8cdf2 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Received unexpected event network-vif-unplugged-0576b708-1b9b-4159-8079-9b97d8b4e70d for instance with vm_state shelved and task_state shelving_offloading. [ 897.618028] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 897.619022] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de2e3cc-d58c-4160-a29c-cf464246ca21 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.627792] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.628087] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39202241-b267-4936-9146-d32e3b7988cd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.709246] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.719695] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.719977] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.720217] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleting the datastore file [datastore2] ad585ebb-2072-45df-b645-94c9fa93576b {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.720583] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a282dce-123c-47af-b502-4203c74dfe86 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.731433] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 897.731433] env[63293]: value = "task-1327831" [ 897.731433] env[63293]: _type = "Task" [ 897.731433] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.742012] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327831, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.891905] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96811118-c7bb-4111-953d-4adee189b2ef {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.902339] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da07e13-290d-45ed-8863-80851b4d8c16 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.937630] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefae532-cf9d-4e03-9532-78f49fcb1b70 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.946390] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b45f1b0-24da-4c08-a05a-53883f12b6a9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.961426] env[63293]: DEBUG nova.compute.provider_tree [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.986724] env[63293]: DEBUG nova.network.neutron [-] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.209155] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.242573] env[63293]: DEBUG oslo_vmware.api [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327831, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.300647} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.242877] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.243084] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 898.243272] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 898.264113] env[63293]: INFO nova.scheduler.client.report [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleted allocations for instance ad585ebb-2072-45df-b645-94c9fa93576b [ 898.464841] env[63293]: DEBUG nova.scheduler.client.report [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.489767] env[63293]: INFO nova.compute.manager [-] [instance: 6cd8b0eb-4bfe-49e0-b392-900f50865ad9] Took 1.27 seconds to deallocate network for instance. [ 898.709288] env[63293]: DEBUG oslo_vmware.api [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Task: {'id': task-1327825, 'name': ReconfigVM_Task, 'duration_secs': 5.914258} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.710699] env[63293]: DEBUG oslo_concurrency.lockutils [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] Releasing lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.710699] env[63293]: DEBUG nova.virt.vmwareapi.vmops [req-f952ba30-879f-4fa7-8ecb-f6d7c2f90260 req-a5101db3-19e6-4ddc-afd9-ed60a7247061 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Reconfigured VM to detach interface {{(pid=63293) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 898.710699] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.117s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.710699] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.710924] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.710971] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.713126] env[63293]: INFO nova.compute.manager [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Terminating instance [ 898.715058] env[63293]: DEBUG nova.compute.manager [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 898.715298] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 898.716138] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252c46ca-9f3c-4026-ba76-98d62f651396 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.724759] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 898.725173] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45f171fe-c594-4aa7-b5fc-65c2e6981353 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.733236] env[63293]: DEBUG oslo_vmware.api [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 898.733236] env[63293]: value = "task-1327832" [ 898.733236] env[63293]: _type = "Task" [ 898.733236] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.742615] env[63293]: DEBUG oslo_vmware.api [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327832, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.769165] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.970724] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.971200] env[63293]: DEBUG nova.compute.manager [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 898.974438] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.842s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.974688] env[63293]: DEBUG nova.objects.instance [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63293) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 899.246825] env[63293]: DEBUG oslo_vmware.api [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327832, 'name': PowerOffVM_Task, 'duration_secs': 0.216202} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.247135] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 899.247314] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 899.247590] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47300d55-fcb2-4c67-a44c-21d3a345f5fa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.330792] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 899.331132] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 899.331367] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Deleting the datastore file [datastore1] 65c2f72d-6bbe-45ac-8efc-401e0664390f {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.331678] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1874d4f5-5f45-4fc8-b98e-6df2bd102780 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.340450] env[63293]: DEBUG oslo_vmware.api [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 899.340450] env[63293]: value = "task-1327834" [ 899.340450] env[63293]: _type = "Task" [ 899.340450] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.349339] env[63293]: DEBUG oslo_vmware.api [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327834, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.480052] env[63293]: DEBUG nova.compute.utils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.484532] env[63293]: DEBUG nova.compute.manager [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 899.484897] env[63293]: DEBUG nova.network.neutron [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 899.516323] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf23c5-47a1-4b33-ba7f-9c6e9b441754 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "6cd8b0eb-4bfe-49e0-b392-900f50865ad9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.356s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.550860] env[63293]: DEBUG nova.policy [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b94ec4d2961a43ca9a331a6051fa774d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7744835de0bc4e30b1b29ee5a439cd0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 899.753214] env[63293]: DEBUG nova.compute.manager [req-d19b63e3-f39f-41ce-9142-d7477390ebe7 req-c0bb4dd6-48c7-4ff8-b446-cf279186c155 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Received event network-changed-0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.753539] env[63293]: DEBUG nova.compute.manager [req-d19b63e3-f39f-41ce-9142-d7477390ebe7 req-c0bb4dd6-48c7-4ff8-b446-cf279186c155 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Refreshing instance network info cache due to event network-changed-0576b708-1b9b-4159-8079-9b97d8b4e70d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 899.753798] env[63293]: DEBUG oslo_concurrency.lockutils [req-d19b63e3-f39f-41ce-9142-d7477390ebe7 req-c0bb4dd6-48c7-4ff8-b446-cf279186c155 service nova] Acquiring lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.753980] env[63293]: DEBUG oslo_concurrency.lockutils [req-d19b63e3-f39f-41ce-9142-d7477390ebe7 req-c0bb4dd6-48c7-4ff8-b446-cf279186c155 service nova] Acquired lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.754105] env[63293]: DEBUG nova.network.neutron [req-d19b63e3-f39f-41ce-9142-d7477390ebe7 req-c0bb4dd6-48c7-4ff8-b446-cf279186c155 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Refreshing network info cache for port 0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 899.851406] env[63293]: DEBUG oslo_vmware.api [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327834, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.310331} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.852148] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 899.852148] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 899.852148] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 899.852400] env[63293]: INFO nova.compute.manager [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 899.852654] env[63293]: DEBUG oslo.service.loopingcall [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.853387] env[63293]: DEBUG nova.network.neutron [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Successfully created port: 7918ba6d-5544-486f-a409-ab8077d44efc {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 899.855343] env[63293]: DEBUG nova.compute.manager [-] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 899.855448] env[63293]: DEBUG nova.network.neutron [-] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 899.921333] env[63293]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port e777ae42-3549-4573-866c-8353417608f1 could not be found.", "detail": ""}} {{(pid=63293) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 899.921569] env[63293]: DEBUG nova.network.neutron [-] Unable to show port e777ae42-3549-4573-866c-8353417608f1 as it no longer exists. {{(pid=63293) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 899.990009] env[63293]: DEBUG nova.compute.manager [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 899.996351] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5c0e9749-df71-462d-8fb8-6a1725cbf3c3 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.022s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.998345] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.503s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.367189] env[63293]: DEBUG oslo_vmware.rw_handles [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f23633-5d1b-4574-1fda-221cdc3fc9a2/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 900.368304] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-096c929f-db76-4c6e-b745-99cbfc402147 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.376390] env[63293]: DEBUG oslo_vmware.rw_handles [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f23633-5d1b-4574-1fda-221cdc3fc9a2/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 900.376698] env[63293]: ERROR oslo_vmware.rw_handles [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f23633-5d1b-4574-1fda-221cdc3fc9a2/disk-0.vmdk due to incomplete transfer. [ 900.377024] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c77d7be4-303c-4328-83ae-9d64e6c69881 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.388147] env[63293]: DEBUG oslo_vmware.rw_handles [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f23633-5d1b-4574-1fda-221cdc3fc9a2/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 900.388454] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Uploaded image 8b176c14-05dc-43c1-a475-57283dc96350 to the Glance image server {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 900.393232] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Destroying the VM {{(pid=63293) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 900.393232] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b26f0ca0-8422-4a58-ac59-f3a3243d275a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.401880] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 900.401880] env[63293]: value = "task-1327835" [ 900.401880] env[63293]: _type = "Task" [ 900.401880] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.415175] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327835, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.701118] env[63293]: DEBUG nova.network.neutron [req-d19b63e3-f39f-41ce-9142-d7477390ebe7 req-c0bb4dd6-48c7-4ff8-b446-cf279186c155 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updated VIF entry in instance network info cache for port 0576b708-1b9b-4159-8079-9b97d8b4e70d. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 900.702632] env[63293]: DEBUG nova.network.neutron [req-d19b63e3-f39f-41ce-9142-d7477390ebe7 req-c0bb4dd6-48c7-4ff8-b446-cf279186c155 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updating instance_info_cache with network_info: [{"id": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "address": "fa:16:3e:c0:e8:15", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap0576b708-1b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.853355] env[63293]: DEBUG nova.compute.manager [req-96810e52-2191-47c3-85f3-05fbc5f3c0c1 req-e237cc77-8418-4f6d-ac2e-a5faf8c89019 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Received event network-vif-deleted-a34b2cc4-afe1-463c-bb22-470ea54d8789 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.853609] env[63293]: INFO nova.compute.manager [req-96810e52-2191-47c3-85f3-05fbc5f3c0c1 req-e237cc77-8418-4f6d-ac2e-a5faf8c89019 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Neutron deleted interface a34b2cc4-afe1-463c-bb22-470ea54d8789; detaching it from the instance and deleting it from the info cache [ 900.853876] env[63293]: DEBUG nova.network.neutron [req-96810e52-2191-47c3-85f3-05fbc5f3c0c1 req-e237cc77-8418-4f6d-ac2e-a5faf8c89019 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.908580] env[63293]: DEBUG nova.network.neutron [-] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.915822] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327835, 'name': Destroy_Task, 'duration_secs': 0.468762} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.916116] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Destroyed the VM [ 900.916484] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Deleting Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 900.916615] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c7d0c3a8-a6c0-4d94-821b-1a3a2adea93d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.924540] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 900.924540] env[63293]: value = "task-1327836" [ 900.924540] env[63293]: _type = "Task" [ 900.924540] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.941016] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327836, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.015114] env[63293]: DEBUG nova.compute.manager [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 901.021151] env[63293]: INFO nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating resource usage from migration 5484fbb9-cb45-4302-9128-7eba0d080fc7 [ 901.038510] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 901.038641] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 65c2f72d-6bbe-45ac-8efc-401e0664390f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 901.040162] env[63293]: WARNING nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 831d7d9b-364f-4c29-bda5-e6c38291a973 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 901.040162] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 9ddf4f23-1279-4cbd-8212-10f344060445 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 901.040162] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance d753598d-a92a-4515-9ad1-d386294f7a99 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 901.040162] env[63293]: WARNING nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 828a1165-3829-431c-9edb-3a5d3a8054ec is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 901.040162] env[63293]: WARNING nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance a15fc833-d9e8-4340-ab4f-c4a0b2d0e703 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 901.040162] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 901.040162] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance eec9bb88-c5c4-4d8a-8a30-84b8d89e153d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 901.040162] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 901.040162] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Migration 5484fbb9-cb45-4302-9128-7eba0d080fc7 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 901.040162] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 76ebe2ec-332c-4d3a-957e-d0c928dafdc7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 901.057020] env[63293]: DEBUG nova.virt.hardware [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.057384] env[63293]: DEBUG nova.virt.hardware [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.057652] env[63293]: DEBUG nova.virt.hardware [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.057878] env[63293]: DEBUG nova.virt.hardware [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.058080] env[63293]: DEBUG nova.virt.hardware [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.058274] env[63293]: DEBUG nova.virt.hardware [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.058674] env[63293]: DEBUG nova.virt.hardware [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.058862] env[63293]: DEBUG nova.virt.hardware [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.059107] env[63293]: DEBUG nova.virt.hardware [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.059380] env[63293]: DEBUG nova.virt.hardware [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.059706] env[63293]: DEBUG nova.virt.hardware [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.061224] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6691aacc-5b50-499b-89c0-acccf3953bdd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.070245] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2af40c-3735-40a5-aea3-dab5ddc849d2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.205887] env[63293]: DEBUG oslo_concurrency.lockutils [req-d19b63e3-f39f-41ce-9142-d7477390ebe7 req-c0bb4dd6-48c7-4ff8-b446-cf279186c155 service nova] Releasing lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.360779] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41e88a0b-66a7-4596-81a5-539b82ae62f2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.374873] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1635fe8-fa7f-428e-b1c7-aa9231de2655 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.404558] env[63293]: DEBUG nova.network.neutron [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Successfully updated port: 7918ba6d-5544-486f-a409-ab8077d44efc {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 901.414538] env[63293]: INFO nova.compute.manager [-] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Took 1.56 seconds to deallocate network for instance. [ 901.414992] env[63293]: DEBUG nova.compute.manager [req-96810e52-2191-47c3-85f3-05fbc5f3c0c1 req-e237cc77-8418-4f6d-ac2e-a5faf8c89019 service nova] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Detach interface failed, port_id=a34b2cc4-afe1-463c-bb22-470ea54d8789, reason: Instance 65c2f72d-6bbe-45ac-8efc-401e0664390f could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 901.435818] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327836, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.492042] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.492313] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.543958] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 5b1264b4-2554-4c62-9ff9-4616b03d3609 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 901.670035] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "ad585ebb-2072-45df-b645-94c9fa93576b" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.836133] env[63293]: DEBUG nova.compute.manager [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Received event network-vif-plugged-7918ba6d-5544-486f-a409-ab8077d44efc {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.836434] env[63293]: DEBUG oslo_concurrency.lockutils [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] Acquiring lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.836835] env[63293]: DEBUG oslo_concurrency.lockutils [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] Lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.837030] env[63293]: DEBUG oslo_concurrency.lockutils [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] Lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.837411] env[63293]: DEBUG nova.compute.manager [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] No waiting events found dispatching network-vif-plugged-7918ba6d-5544-486f-a409-ab8077d44efc {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 901.837696] env[63293]: WARNING nova.compute.manager [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Received unexpected event network-vif-plugged-7918ba6d-5544-486f-a409-ab8077d44efc for instance with vm_state building and task_state spawning. [ 901.838000] env[63293]: DEBUG nova.compute.manager [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Received event network-changed-7918ba6d-5544-486f-a409-ab8077d44efc {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.838187] env[63293]: DEBUG nova.compute.manager [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Refreshing instance network info cache due to event network-changed-7918ba6d-5544-486f-a409-ab8077d44efc. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 901.838419] env[63293]: DEBUG oslo_concurrency.lockutils [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] Acquiring lock "refresh_cache-5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.838530] env[63293]: DEBUG oslo_concurrency.lockutils [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] Acquired lock "refresh_cache-5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.838725] env[63293]: DEBUG nova.network.neutron [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Refreshing network info cache for port 7918ba6d-5544-486f-a409-ab8077d44efc {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.916805] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "refresh_cache-5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.923392] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.939257] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327836, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.944073] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "2557bded-6ccb-4c0f-8b97-34529b4bb740" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.944339] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "2557bded-6ccb-4c0f-8b97-34529b4bb740" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.995125] env[63293]: DEBUG nova.compute.manager [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.046501] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 8625a358-ec48-46b9-8738-c49a2ba58362 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 902.046718] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (619725c3-6194-4724-94b1-1dd499be55d0): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 902.381711] env[63293]: DEBUG nova.network.neutron [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 902.441804] env[63293]: DEBUG oslo_vmware.api [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327836, 'name': RemoveSnapshot_Task, 'duration_secs': 1.039584} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.445481] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Deleted Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 902.445853] env[63293]: INFO nova.compute.manager [None req-08e790f6-73ae-4997-bed8-490f30b7526c tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Took 15.15 seconds to snapshot the instance on the hypervisor. [ 902.450274] env[63293]: DEBUG nova.compute.manager [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.502994] env[63293]: DEBUG nova.network.neutron [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.521650] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.549958] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance fe55a9b7-28b5-4b04-84f1-6b85b579a144 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 902.550127] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 902.550325] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 902.579744] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b7c635-225c-5bdf-4563-bfbf0e7beeb4/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 902.581116] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11146d79-5878-4a0e-b49f-995c572edd10 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.590160] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b7c635-225c-5bdf-4563-bfbf0e7beeb4/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 902.590338] env[63293]: ERROR oslo_vmware.rw_handles [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b7c635-225c-5bdf-4563-bfbf0e7beeb4/disk-0.vmdk due to incomplete transfer. [ 902.590581] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-307d2e80-8564-45e6-9fb5-1bf87d1459c8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.597849] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b7c635-225c-5bdf-4563-bfbf0e7beeb4/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 902.598068] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Uploaded image 43bec244-5978-48cb-96cf-849335c88aa8 to the Glance image server {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 902.600349] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Destroying the VM {{(pid=63293) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 902.600631] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d3524d97-827b-4473-abaa-ea5a985491af {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.607935] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 902.607935] env[63293]: value = "task-1327837" [ 902.607935] env[63293]: _type = "Task" [ 902.607935] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.616367] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327837, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.762761] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9562da9-0654-45f6-9aef-3cfe4d9ae89a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.770868] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568db68d-450a-4e5c-96c9-5fe6e4537659 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.800859] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84aabd3-c983-4ffd-b8e7-169db7cdc2d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.808323] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20ef0d9-1d91-47d7-8e49-680567beb895 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.821445] env[63293]: DEBUG nova.compute.provider_tree [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.972036] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.006293] env[63293]: DEBUG oslo_concurrency.lockutils [req-10d8476a-c756-433b-b3a0-c059cee3d2fb req-701d3997-15d6-4101-ad0a-26c480248369 service nova] Releasing lock "refresh_cache-5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.006711] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "refresh_cache-5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.006874] env[63293]: DEBUG nova.network.neutron [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 903.118230] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327837, 'name': Destroy_Task, 'duration_secs': 0.320145} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.118514] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Destroyed the VM [ 903.118761] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Deleting Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 903.119023] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a9f03505-0ec9-4ae9-af0f-899d8bb09fd0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.126547] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 903.126547] env[63293]: value = "task-1327838" [ 903.126547] env[63293]: _type = "Task" [ 903.126547] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.135440] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327838, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.324249] env[63293]: DEBUG nova.scheduler.client.report [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.538961] env[63293]: DEBUG nova.network.neutron [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 903.642132] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327838, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.676574] env[63293]: DEBUG nova.network.neutron [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Updating instance_info_cache with network_info: [{"id": "7918ba6d-5544-486f-a409-ab8077d44efc", "address": "fa:16:3e:f6:3b:a4", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7918ba6d-55", "ovs_interfaceid": "7918ba6d-5544-486f-a409-ab8077d44efc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.830081] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 903.830329] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.832s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.830624] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.484s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.831166] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.833372] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.480s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.833571] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.835287] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.351s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.835475] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.836981] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 14.249s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.865185] env[63293]: INFO nova.scheduler.client.report [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Deleted allocations for instance 831d7d9b-364f-4c29-bda5-e6c38291a973 [ 903.866904] env[63293]: INFO nova.scheduler.client.report [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Deleted allocations for instance a15fc833-d9e8-4340-ab4f-c4a0b2d0e703 [ 903.878905] env[63293]: INFO nova.scheduler.client.report [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Deleted allocations for instance 828a1165-3829-431c-9edb-3a5d3a8054ec [ 904.136882] env[63293]: DEBUG oslo_vmware.api [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327838, 'name': RemoveSnapshot_Task, 'duration_secs': 0.515813} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.137189] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Deleted Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 904.137507] env[63293]: INFO nova.compute.manager [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Took 14.41 seconds to snapshot the instance on the hypervisor. [ 904.178903] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "refresh_cache-5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.179268] env[63293]: DEBUG nova.compute.manager [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Instance network_info: |[{"id": "7918ba6d-5544-486f-a409-ab8077d44efc", "address": "fa:16:3e:f6:3b:a4", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7918ba6d-55", "ovs_interfaceid": "7918ba6d-5544-486f-a409-ab8077d44efc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 904.179732] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:3b:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ef746c57-cd18-4883-a0e9-c52937aaf41d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7918ba6d-5544-486f-a409-ab8077d44efc', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 904.188643] env[63293]: DEBUG oslo.service.loopingcall [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.192022] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 904.192022] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7fe3e4d2-c0f9-461a-bf4f-155a853af08d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.209888] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.209888] env[63293]: value = "task-1327839" [ 904.209888] env[63293]: _type = "Task" [ 904.209888] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.217561] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327839, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.343295] env[63293]: INFO nova.compute.claims [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.384492] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8c580e15-6ac3-456e-b527-284628dfa8d5 tempest-AttachInterfacesUnderV243Test-202663734 tempest-AttachInterfacesUnderV243Test-202663734-project-member] Lock "831d7d9b-364f-4c29-bda5-e6c38291a973" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.623s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.387853] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d490367e-1b30-403b-8e94-b1eccfe2ec70 tempest-InstanceActionsTestJSON-1480768882 tempest-InstanceActionsTestJSON-1480768882-project-member] Lock "a15fc833-d9e8-4340-ab4f-c4a0b2d0e703" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.366s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.391294] env[63293]: DEBUG oslo_concurrency.lockutils [None req-514b9235-ce06-4dcb-9229-0b77303836e5 tempest-ServerShowV257Test-1389303658 tempest-ServerShowV257Test-1389303658-project-member] Lock "828a1165-3829-431c-9edb-3a5d3a8054ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.860s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.710322] env[63293]: DEBUG nova.compute.manager [None req-c57e88d5-d3bf-4153-9ddc-df19acb1e1b9 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Found 1 images (rotation: 2) {{(pid=63293) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 904.722070] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327839, 'name': CreateVM_Task, 'duration_secs': 0.348147} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.722070] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 904.723477] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.724361] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.725197] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.725513] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e1842ac-91a4-44e5-ab32-1ecb0bdf5a5f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.731715] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 904.731715] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522b4097-a610-c652-fd29-af2211dd82b8" [ 904.731715] env[63293]: _type = "Task" [ 904.731715] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.743341] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522b4097-a610-c652-fd29-af2211dd82b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.850145] env[63293]: INFO nova.compute.resource_tracker [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating resource usage from migration 5484fbb9-cb45-4302-9128-7eba0d080fc7 [ 905.081278] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac8b093-8a99-4356-95c0-8bfc45e6cf2d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.090749] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e5305a-51a8-48e0-a8ea-a208cd6148d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.131174] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab67b019-0a50-4384-821b-927d5acdee03 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.139939] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579d029f-ac49-433d-a87b-b6a1f50b2698 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.154972] env[63293]: DEBUG nova.compute.provider_tree [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.245953] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522b4097-a610-c652-fd29-af2211dd82b8, 'name': SearchDatastore_Task, 'duration_secs': 0.010655} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.246336] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.246580] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.246825] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.246978] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.247193] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.247473] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3146901a-8ea2-4836-b42b-2931a4f5100a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.259774] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.259774] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 905.259774] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10166da1-0290-4a96-b64b-7e4291a5fd77 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.264151] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 905.264151] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]521582f8-c794-5ddc-7031-880e78b29ddd" [ 905.264151] env[63293]: _type = "Task" [ 905.264151] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.271897] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]521582f8-c794-5ddc-7031-880e78b29ddd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.601620] env[63293]: DEBUG nova.compute.manager [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.602577] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e9c684-3db6-4b3e-b064-5a45400061b0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.659187] env[63293]: DEBUG nova.scheduler.client.report [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.777606] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]521582f8-c794-5ddc-7031-880e78b29ddd, 'name': SearchDatastore_Task, 'duration_secs': 0.009237} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.778422] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-187aec61-caf8-4996-b19b-646c4e00665f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.783776] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 905.783776] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52187489-7b74-cf9e-a43d-596dff684fcc" [ 905.783776] env[63293]: _type = "Task" [ 905.783776] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.791549] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52187489-7b74-cf9e-a43d-596dff684fcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.116069] env[63293]: INFO nova.compute.manager [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] instance snapshotting [ 906.117021] env[63293]: DEBUG nova.objects.instance [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'flavor' on Instance uuid 9ddf4f23-1279-4cbd-8212-10f344060445 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.172019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.333s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.172019] env[63293]: INFO nova.compute.manager [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Migrating [ 906.172019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.172019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "compute-rpcapi-router" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.178136] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.874s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.178136] env[63293]: INFO nova.compute.claims [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.180222] env[63293]: INFO nova.compute.rpcapi [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 906.181115] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "compute-rpcapi-router" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.295846] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52187489-7b74-cf9e-a43d-596dff684fcc, 'name': SearchDatastore_Task, 'duration_secs': 0.011072} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.295846] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.295846] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d/5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 906.296193] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c8ff340a-0bbd-40c3-8bc6-1762f4acb1cd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.304721] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 906.304721] env[63293]: value = "task-1327840" [ 906.304721] env[63293]: _type = "Task" [ 906.304721] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.312901] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327840, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.635215] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9316bf4c-2e5f-455e-ada3-0054062e4367 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.653992] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71735d33-dbec-4007-a08a-00d5c45ea034 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.700270] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.700270] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.700270] env[63293]: DEBUG nova.network.neutron [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 906.816215] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327840, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.167066] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Creating Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 907.167066] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e485d139-f581-4689-8239-1792677c6e0d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.174918] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 907.174918] env[63293]: value = "task-1327841" [ 907.174918] env[63293]: _type = "Task" [ 907.174918] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.184693] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327841, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.317692] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327840, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526171} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.319983] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d/5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 907.320269] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 907.320584] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-89811c0b-cf8a-4a7e-82da-482dd63130cd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.329730] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 907.329730] env[63293]: value = "task-1327842" [ 907.329730] env[63293]: _type = "Task" [ 907.329730] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.339443] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327842, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.404377] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200a493c-2f19-43a7-bfec-09b2f7526995 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.416029] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f10d30b-8c06-40ef-af89-5d6421b98573 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.454974] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288caef8-bfea-49e8-a2b3-3004e7c1e0df {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.462873] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656f9d35-41bf-4e16-a8c2-a6febdb91c4d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.476926] env[63293]: DEBUG nova.compute.provider_tree [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.500143] env[63293]: DEBUG nova.network.neutron [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance_info_cache with network_info: [{"id": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "address": "fa:16:3e:12:25:eb", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46d5563f-41", "ovs_interfaceid": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.690333] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327841, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.842138] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327842, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07567} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.842138] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.843116] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3be208-4979-4a57-b333-98c528e49ddf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.867756] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d/5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.869412] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b15c85a-474a-4578-87ef-c819ea1f00f7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.888466] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 907.888466] env[63293]: value = "task-1327843" [ 907.888466] env[63293]: _type = "Task" [ 907.888466] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.897657] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327843, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.982018] env[63293]: DEBUG nova.scheduler.client.report [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.004737] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.187635] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327841, 'name': CreateSnapshot_Task, 'duration_secs': 0.518225} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.188489] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Created Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 908.189773] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d159421-5a59-44d9-b597-f14e6eae5201 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.304941] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.304941] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.402022] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327843, 'name': ReconfigVM_Task, 'duration_secs': 0.36028} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.402022] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d/5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.402022] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e69f322-049d-4961-9daf-bb149051130e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.410019] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 908.410019] env[63293]: value = "task-1327844" [ 908.410019] env[63293]: _type = "Task" [ 908.410019] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.417746] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327844, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.485971] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.486924] env[63293]: DEBUG nova.compute.manager [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.490993] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.435s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.496530] env[63293]: INFO nova.compute.claims [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.715877] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Creating linked-clone VM from snapshot {{(pid=63293) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 908.717361] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5bf2bdb2-f992-4bf6-8de4-15a0c765e628 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.728992] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 908.728992] env[63293]: value = "task-1327845" [ 908.728992] env[63293]: _type = "Task" [ 908.728992] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.737542] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327845, 'name': CloneVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.807875] env[63293]: INFO nova.compute.manager [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Detaching volume 981997fd-c500-4d22-bd91-77b5fb478502 [ 908.844960] env[63293]: INFO nova.virt.block_device [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Attempting to driver detach volume 981997fd-c500-4d22-bd91-77b5fb478502 from mountpoint /dev/sdb [ 908.845250] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Volume detach. Driver type: vmdk {{(pid=63293) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 908.845474] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283762', 'volume_id': '981997fd-c500-4d22-bd91-77b5fb478502', 'name': 'volume-981997fd-c500-4d22-bd91-77b5fb478502', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0977c79a-2ef3-4ec9-a0a0-de1ed5799723', 'attached_at': '', 'detached_at': '', 'volume_id': '981997fd-c500-4d22-bd91-77b5fb478502', 'serial': '981997fd-c500-4d22-bd91-77b5fb478502'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 908.847073] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f90368-531c-4498-9a27-da83278a4db1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.871442] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ea48a3-0d00-4d06-97e4-1edde52a87b0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.881429] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d044caa-c598-4c2f-9c4b-cb5e9da64a2f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.915305] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-479009e4-e940-4b12-9522-29c63fe94ee1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.920445] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "d753598d-a92a-4515-9ad1-d386294f7a99" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.920445] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.938160] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] The volume has not been displaced from its original location: [datastore2] volume-981997fd-c500-4d22-bd91-77b5fb478502/volume-981997fd-c500-4d22-bd91-77b5fb478502.vmdk. No consolidation needed. {{(pid=63293) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 908.944079] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Reconfiguring VM instance instance-00000039 to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 908.944466] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327844, 'name': Rename_Task, 'duration_secs': 0.141519} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.945732] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e34adbb-791a-4a91-b0b7-d88c4df1575d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.958824] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 908.959913] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c1a4f0b-ee74-47fe-ac31-ff291f201daa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.966734] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 908.966734] env[63293]: value = "task-1327846" [ 908.966734] env[63293]: _type = "Task" [ 908.966734] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.968463] env[63293]: DEBUG oslo_vmware.api [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 908.968463] env[63293]: value = "task-1327847" [ 908.968463] env[63293]: _type = "Task" [ 908.968463] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.979771] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327846, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.983170] env[63293]: DEBUG oslo_vmware.api [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327847, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.004177] env[63293]: DEBUG nova.compute.utils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.008697] env[63293]: DEBUG nova.compute.manager [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 909.008697] env[63293]: DEBUG nova.network.neutron [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 909.081371] env[63293]: DEBUG nova.policy [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4098c09e1c994cbbbb578fa6a3a5a002', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1cf310a9134147caa91f6aec4ebde76e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 909.241442] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327845, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.446470] env[63293]: DEBUG nova.compute.utils [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.485352] env[63293]: DEBUG oslo_vmware.api [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327847, 'name': ReconfigVM_Task, 'duration_secs': 0.391181} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.490052] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Reconfigured VM instance instance-00000039 to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 909.495333] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327846, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.495517] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2be216a9-82b7-4197-a2c6-179fda2cc3bd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.515896] env[63293]: DEBUG nova.network.neutron [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Successfully created port: f6501e07-0e63-4597-86b9-bbe243fabaa3 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.521167] env[63293]: DEBUG nova.compute.manager [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.529694] env[63293]: DEBUG oslo_vmware.api [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 909.529694] env[63293]: value = "task-1327848" [ 909.529694] env[63293]: _type = "Task" [ 909.529694] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.531616] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa755c9-68b1-4156-af15-7d5d6859489f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.543473] env[63293]: DEBUG oslo_vmware.api [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327848, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.560438] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance '76ebe2ec-332c-4d3a-957e-d0c928dafdc7' progress to 0 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 909.743937] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327845, 'name': CloneVM_Task} progress is 95%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.822264] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c4fde3-a12a-4386-a5a3-176294a51dc4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.830028] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d961c7-eaa4-4bf1-8f0a-19575825684f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.865262] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a6e04c-d4d7-4190-8c84-a58e90745dd1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.871978] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b03712-7bf9-4f20-a487-749357525038 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.885844] env[63293]: DEBUG nova.compute.provider_tree [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.951270] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.031s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.978883] env[63293]: DEBUG oslo_vmware.api [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327846, 'name': PowerOnVM_Task, 'duration_secs': 0.578812} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.979180] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.979387] env[63293]: INFO nova.compute.manager [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Took 8.96 seconds to spawn the instance on the hypervisor. [ 909.979676] env[63293]: DEBUG nova.compute.manager [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.980461] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e7c0015-3f4e-4fba-9bf1-c67e87494fb1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.041076] env[63293]: DEBUG oslo_vmware.api [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327848, 'name': ReconfigVM_Task, 'duration_secs': 0.237452} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.041558] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283762', 'volume_id': '981997fd-c500-4d22-bd91-77b5fb478502', 'name': 'volume-981997fd-c500-4d22-bd91-77b5fb478502', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0977c79a-2ef3-4ec9-a0a0-de1ed5799723', 'attached_at': '', 'detached_at': '', 'volume_id': '981997fd-c500-4d22-bd91-77b5fb478502', 'serial': '981997fd-c500-4d22-bd91-77b5fb478502'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 910.069921] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.069921] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94501ca1-52aa-4d97-99e7-190db0dade7a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.076941] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 910.076941] env[63293]: value = "task-1327849" [ 910.076941] env[63293]: _type = "Task" [ 910.076941] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.087774] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327849, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.241558] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327845, 'name': CloneVM_Task, 'duration_secs': 1.175481} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.243781] env[63293]: INFO nova.virt.vmwareapi.vmops [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Created linked-clone VM from snapshot [ 910.243781] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a13d76f-93e4-4f9f-b664-8943f901f845 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.252310] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Uploading image 309b9adc-458f-4eaa-8a37-2f7ffc94d5dc {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 910.284640] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 910.284640] env[63293]: value = "vm-283789" [ 910.284640] env[63293]: _type = "VirtualMachine" [ 910.284640] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 910.284937] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c4880262-9c88-443d-a965-5f2c59ce1575 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.293034] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lease: (returnval){ [ 910.293034] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523ac93a-fca2-a77f-ee6c-20f1525bf66c" [ 910.293034] env[63293]: _type = "HttpNfcLease" [ 910.293034] env[63293]: } obtained for exporting VM: (result){ [ 910.293034] env[63293]: value = "vm-283789" [ 910.293034] env[63293]: _type = "VirtualMachine" [ 910.293034] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 910.293408] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the lease: (returnval){ [ 910.293408] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523ac93a-fca2-a77f-ee6c-20f1525bf66c" [ 910.293408] env[63293]: _type = "HttpNfcLease" [ 910.293408] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 910.300673] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 910.300673] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523ac93a-fca2-a77f-ee6c-20f1525bf66c" [ 910.300673] env[63293]: _type = "HttpNfcLease" [ 910.300673] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 910.396653] env[63293]: DEBUG nova.scheduler.client.report [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.508738] env[63293]: INFO nova.compute.manager [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Took 29.09 seconds to build instance. [ 910.537035] env[63293]: DEBUG nova.compute.manager [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.565738] env[63293]: DEBUG nova.virt.hardware [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.566018] env[63293]: DEBUG nova.virt.hardware [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.566183] env[63293]: DEBUG nova.virt.hardware [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.566372] env[63293]: DEBUG nova.virt.hardware [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.566528] env[63293]: DEBUG nova.virt.hardware [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.566702] env[63293]: DEBUG nova.virt.hardware [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.568616] env[63293]: DEBUG nova.virt.hardware [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.568616] env[63293]: DEBUG nova.virt.hardware [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.568616] env[63293]: DEBUG nova.virt.hardware [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.568616] env[63293]: DEBUG nova.virt.hardware [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.568616] env[63293]: DEBUG nova.virt.hardware [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.569708] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cae56db-a9e8-4338-90cc-87dae49f8586 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.581529] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd86d18-ab8d-4aa3-8177-a179521fb51c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.591043] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327849, 'name': PowerOffVM_Task, 'duration_secs': 0.206063} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.599074] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.599363] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance '76ebe2ec-332c-4d3a-957e-d0c928dafdc7' progress to 17 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 910.622030] env[63293]: DEBUG nova.objects.instance [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'flavor' on Instance uuid 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.802174] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 910.802174] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523ac93a-fca2-a77f-ee6c-20f1525bf66c" [ 910.802174] env[63293]: _type = "HttpNfcLease" [ 910.802174] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 910.802511] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 910.802511] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523ac93a-fca2-a77f-ee6c-20f1525bf66c" [ 910.802511] env[63293]: _type = "HttpNfcLease" [ 910.802511] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 910.803242] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2609342e-3f42-4451-8568-48d99edbbf9f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.810910] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52017786-2c19-5c20-2ae3-52b86b0c3804/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 910.811716] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52017786-2c19-5c20-2ae3-52b86b0c3804/disk-0.vmdk for reading. {{(pid=63293) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 910.903203] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.903747] env[63293]: DEBUG nova.compute.manager [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 910.906523] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.138s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.909788] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.909788] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.986s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.909788] env[63293]: DEBUG nova.objects.instance [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'resources' on Instance uuid 65c2f72d-6bbe-45ac-8efc-401e0664390f {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.952216] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-668e6dd3-36e0-42a7-818d-38dfc99cb6de {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.987916] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "d753598d-a92a-4515-9ad1-d386294f7a99" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.988982] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.989591] env[63293]: INFO nova.compute.manager [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Attaching volume 6c9dda44-62d5-4ee9-8961-4b9fec04802a to /dev/sdb [ 911.019259] env[63293]: DEBUG oslo_concurrency.lockutils [None req-70abd144-f0ed-4653-95b1-8e05ed6258a1 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.127s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.047565] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad253c0-4722-4a07-b857-d664449bcb3f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.055347] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d086c26-71f9-488a-926a-c4cddbb177cb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.075944] env[63293]: DEBUG nova.virt.block_device [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Updating existing volume attachment record: 231d4292-9d38-4ac9-bfe9-935b8e286f6d {{(pid=63293) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 911.105950] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:32:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.106904] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.106904] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.106904] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.106904] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.106904] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.107248] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.107248] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.107395] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.107746] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.107746] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.116105] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccb3746c-6b3c-4cd1-b6e6-ed56e8c4cbfb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.134976] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 911.134976] env[63293]: value = "task-1327851" [ 911.134976] env[63293]: _type = "Task" [ 911.134976] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.142708] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327851, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.163663] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "f429869f-9e9f-443e-a980-67a80c3bd799" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.164385] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "f429869f-9e9f-443e-a980-67a80c3bd799" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.413499] env[63293]: DEBUG nova.compute.utils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.419709] env[63293]: DEBUG nova.compute.manager [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 911.420073] env[63293]: DEBUG nova.network.neutron [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 911.422193] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b19a3f23-b0b7-41ba-8d5c-7e4a77acb2fe tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 34.326s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.423228] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 9.754s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.423472] env[63293]: INFO nova.compute.manager [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Unshelving [ 911.476741] env[63293]: DEBUG nova.policy [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '724f5f1ec5174808adb3ce7395052f44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2bb8a8bec2a04cb292af72d1216bb355', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.492683] env[63293]: DEBUG nova.compute.manager [req-4aeab982-1667-4756-aea8-b869831167a0 req-a10ae2b1-267a-4e5c-95ae-c3d7614e62ad service nova] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Received event network-vif-plugged-f6501e07-0e63-4597-86b9-bbe243fabaa3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.493852] env[63293]: DEBUG oslo_concurrency.lockutils [req-4aeab982-1667-4756-aea8-b869831167a0 req-a10ae2b1-267a-4e5c-95ae-c3d7614e62ad service nova] Acquiring lock "5b1264b4-2554-4c62-9ff9-4616b03d3609-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.494205] env[63293]: DEBUG oslo_concurrency.lockutils [req-4aeab982-1667-4756-aea8-b869831167a0 req-a10ae2b1-267a-4e5c-95ae-c3d7614e62ad service nova] Lock "5b1264b4-2554-4c62-9ff9-4616b03d3609-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.494544] env[63293]: DEBUG oslo_concurrency.lockutils [req-4aeab982-1667-4756-aea8-b869831167a0 req-a10ae2b1-267a-4e5c-95ae-c3d7614e62ad service nova] Lock "5b1264b4-2554-4c62-9ff9-4616b03d3609-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.495197] env[63293]: DEBUG nova.compute.manager [req-4aeab982-1667-4756-aea8-b869831167a0 req-a10ae2b1-267a-4e5c-95ae-c3d7614e62ad service nova] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] No waiting events found dispatching network-vif-plugged-f6501e07-0e63-4597-86b9-bbe243fabaa3 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.495345] env[63293]: WARNING nova.compute.manager [req-4aeab982-1667-4756-aea8-b869831167a0 req-a10ae2b1-267a-4e5c-95ae-c3d7614e62ad service nova] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Received unexpected event network-vif-plugged-f6501e07-0e63-4597-86b9-bbe243fabaa3 for instance with vm_state building and task_state spawning. [ 911.564496] env[63293]: DEBUG oslo_concurrency.lockutils [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.586482] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.586732] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.586973] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.587380] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.587853] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.596729] env[63293]: INFO nova.compute.manager [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Terminating instance [ 911.603540] env[63293]: DEBUG nova.compute.manager [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 911.603540] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 911.604567] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7d16f0-7339-4703-986a-6f6e1a8e40a2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.615886] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 911.616540] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d46a0604-03c1-4fe3-81c0-c227ec74de1c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.628122] env[63293]: DEBUG oslo_vmware.api [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 911.628122] env[63293]: value = "task-1327855" [ 911.628122] env[63293]: _type = "Task" [ 911.628122] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.637898] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d58db60b-7c72-4a3d-bd49-a951ecc2c124 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.334s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.653448] env[63293]: DEBUG oslo_concurrency.lockutils [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.089s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.654219] env[63293]: DEBUG nova.compute.manager [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.658150] env[63293]: DEBUG oslo_vmware.api [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327855, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.658818] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7ff5c7-545c-4b76-87f2-1779e2c2362c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.674845] env[63293]: DEBUG nova.compute.manager [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 911.678676] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327851, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.681932] env[63293]: DEBUG nova.compute.manager [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63293) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 911.683046] env[63293]: DEBUG nova.objects.instance [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'flavor' on Instance uuid 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.726281] env[63293]: DEBUG nova.network.neutron [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Successfully updated port: f6501e07-0e63-4597-86b9-bbe243fabaa3 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.761749] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f195cc9a-d787-44a3-b0cd-724e54410802 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.772487] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0406c57-b11b-4a6e-a42f-4d8ea2f3bd78 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.808481] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbe7de2-6fac-466c-8d97-0378e91b5d55 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.816654] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd702ef-2285-4f78-b3ff-a9e59f6ab2ea {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.831671] env[63293]: DEBUG nova.compute.provider_tree [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.920626] env[63293]: DEBUG nova.compute.manager [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.935471] env[63293]: DEBUG nova.network.neutron [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Successfully created port: 9551ebb5-552d-47d0-a436-994fe2b1d00b {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 912.146704] env[63293]: DEBUG oslo_vmware.api [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327855, 'name': PowerOffVM_Task, 'duration_secs': 0.213903} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.150500] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 912.150938] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 912.151498] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6f133cef-d575-48a7-b55f-35c6dbdab5d3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.160717] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327851, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.193720] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.194434] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52dc6ed6-8c75-4ba5-8e18-8a37a6b8cd11 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.202328] env[63293]: DEBUG oslo_vmware.api [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 912.202328] env[63293]: value = "task-1327857" [ 912.202328] env[63293]: _type = "Task" [ 912.202328] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.209111] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.215241] env[63293]: DEBUG oslo_vmware.api [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327857, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.222906] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 912.223155] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 912.223345] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleting the datastore file [datastore1] 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.223740] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03f95f41-0fdd-47f3-9d42-beb13bb2117d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.229642] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "refresh_cache-5b1264b4-2554-4c62-9ff9-4616b03d3609" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.229917] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquired lock "refresh_cache-5b1264b4-2554-4c62-9ff9-4616b03d3609" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.230408] env[63293]: DEBUG nova.network.neutron [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.233710] env[63293]: DEBUG oslo_vmware.api [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 912.233710] env[63293]: value = "task-1327858" [ 912.233710] env[63293]: _type = "Task" [ 912.233710] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.244964] env[63293]: DEBUG oslo_vmware.api [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327858, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.335878] env[63293]: DEBUG nova.scheduler.client.report [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.433481] env[63293]: DEBUG nova.compute.utils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 912.655663] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327851, 'name': ReconfigVM_Task, 'duration_secs': 1.204358} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.657100] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance '76ebe2ec-332c-4d3a-957e-d0c928dafdc7' progress to 33 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 912.715730] env[63293]: DEBUG oslo_vmware.api [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327857, 'name': PowerOffVM_Task, 'duration_secs': 0.196672} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.716381] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 912.716381] env[63293]: DEBUG nova.compute.manager [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.717254] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb1dde5-96f5-465d-b169-176c55b9e71c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.743695] env[63293]: DEBUG oslo_vmware.api [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327858, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172179} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.744809] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 912.744809] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 912.744809] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 912.744809] env[63293]: INFO nova.compute.manager [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 912.745166] env[63293]: DEBUG oslo.service.loopingcall [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.745357] env[63293]: DEBUG nova.compute.manager [-] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 912.745357] env[63293]: DEBUG nova.network.neutron [-] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 912.778447] env[63293]: DEBUG nova.network.neutron [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 912.845431] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.936s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.847264] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.326s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.848821] env[63293]: INFO nova.compute.claims [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.882836] env[63293]: INFO nova.scheduler.client.report [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Deleted allocations for instance 65c2f72d-6bbe-45ac-8efc-401e0664390f [ 912.930290] env[63293]: DEBUG nova.compute.manager [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 912.936974] env[63293]: INFO nova.virt.block_device [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Booting with volume 658fbd49-f54e-404a-99c2-e7147e4a3700 at /dev/sdb [ 912.962772] env[63293]: DEBUG nova.virt.hardware [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.963458] env[63293]: DEBUG nova.virt.hardware [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.963458] env[63293]: DEBUG nova.virt.hardware [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.963620] env[63293]: DEBUG nova.virt.hardware [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.963662] env[63293]: DEBUG nova.virt.hardware [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.964647] env[63293]: DEBUG nova.virt.hardware [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.964647] env[63293]: DEBUG nova.virt.hardware [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.964647] env[63293]: DEBUG nova.virt.hardware [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.964647] env[63293]: DEBUG nova.virt.hardware [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.964647] env[63293]: DEBUG nova.virt.hardware [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.964989] env[63293]: DEBUG nova.virt.hardware [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.965706] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f114a27b-a809-414b-8d58-ec3a480d6fbd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.976196] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce032eb-92f4-4ba4-88e6-1880858878f8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.980670] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-376d2eec-e1df-427e-bb12-efd53bcf1cdf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.996802] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7c7856-1628-4e64-ba8a-c9b1ee977589 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.013155] env[63293]: DEBUG nova.network.neutron [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Updating instance_info_cache with network_info: [{"id": "f6501e07-0e63-4597-86b9-bbe243fabaa3", "address": "fa:16:3e:06:94:5f", "network": {"id": "02c216f2-c41a-4722-91df-330cd520be3d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f57886847a1446109c6ab061a248dcf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6501e07-0e", "ovs_interfaceid": "f6501e07-0e63-4597-86b9-bbe243fabaa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.029299] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6fd82e0f-59de-402c-9a80-72b8637c4b93 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.039428] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44e23a6-86b1-438b-9fa8-531dbacd4a40 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.055757] env[63293]: DEBUG nova.compute.manager [req-93268fa7-d256-4bbc-bbda-ae66c9def0e7 req-960dbad3-401b-41fa-9df1-6782f0c9404d service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Received event network-vif-deleted-7918ba6d-5544-486f-a409-ab8077d44efc {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.056073] env[63293]: INFO nova.compute.manager [req-93268fa7-d256-4bbc-bbda-ae66c9def0e7 req-960dbad3-401b-41fa-9df1-6782f0c9404d service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Neutron deleted interface 7918ba6d-5544-486f-a409-ab8077d44efc; detaching it from the instance and deleting it from the info cache [ 913.056298] env[63293]: DEBUG nova.network.neutron [req-93268fa7-d256-4bbc-bbda-ae66c9def0e7 req-960dbad3-401b-41fa-9df1-6782f0c9404d service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.068972] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265c7c0f-ca8d-4710-ba89-4d3dddfdc6d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.076776] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcfd8270-d9a3-4d68-a498-c2ccd58332d5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.090164] env[63293]: DEBUG nova.virt.block_device [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updating existing volume attachment record: 36203fea-1792-4438-a344-28a30fa862bb {{(pid=63293) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 913.165126] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 913.165126] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 913.165126] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.165380] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 913.165511] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.165671] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 913.165883] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 913.166062] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 913.166243] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 913.166412] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 913.166589] env[63293]: DEBUG nova.virt.hardware [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.172278] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Reconfiguring VM instance instance-0000004e to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 913.172597] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36eb8acf-252a-48a2-96af-65b213b4ccbf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.190758] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 913.190758] env[63293]: value = "task-1327859" [ 913.190758] env[63293]: _type = "Task" [ 913.190758] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.199256] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327859, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.229095] env[63293]: DEBUG oslo_concurrency.lockutils [None req-dd2be1c7-1596-4526-a51a-87bff20e84fe tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.576s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.397926] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c787e6e0-e55d-4d13-ba99-fd9498cef64c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "65c2f72d-6bbe-45ac-8efc-401e0664390f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.687s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.515466] env[63293]: DEBUG nova.network.neutron [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Successfully updated port: 9551ebb5-552d-47d0-a436-994fe2b1d00b {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.520432] env[63293]: DEBUG nova.compute.manager [req-49868658-dba1-4318-a820-c2f145163118 req-fcab777a-6479-4ebd-b52e-0bcc45cb7f8b service nova] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Received event network-changed-f6501e07-0e63-4597-86b9-bbe243fabaa3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.520432] env[63293]: DEBUG nova.compute.manager [req-49868658-dba1-4318-a820-c2f145163118 req-fcab777a-6479-4ebd-b52e-0bcc45cb7f8b service nova] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Refreshing instance network info cache due to event network-changed-f6501e07-0e63-4597-86b9-bbe243fabaa3. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 913.520432] env[63293]: DEBUG oslo_concurrency.lockutils [req-49868658-dba1-4318-a820-c2f145163118 req-fcab777a-6479-4ebd-b52e-0bcc45cb7f8b service nova] Acquiring lock "refresh_cache-5b1264b4-2554-4c62-9ff9-4616b03d3609" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.530559] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Releasing lock "refresh_cache-5b1264b4-2554-4c62-9ff9-4616b03d3609" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.530559] env[63293]: DEBUG nova.compute.manager [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Instance network_info: |[{"id": "f6501e07-0e63-4597-86b9-bbe243fabaa3", "address": "fa:16:3e:06:94:5f", "network": {"id": "02c216f2-c41a-4722-91df-330cd520be3d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f57886847a1446109c6ab061a248dcf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6501e07-0e", "ovs_interfaceid": "f6501e07-0e63-4597-86b9-bbe243fabaa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 913.531149] env[63293]: DEBUG oslo_concurrency.lockutils [req-49868658-dba1-4318-a820-c2f145163118 req-fcab777a-6479-4ebd-b52e-0bcc45cb7f8b service nova] Acquired lock "refresh_cache-5b1264b4-2554-4c62-9ff9-4616b03d3609" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.531380] env[63293]: DEBUG nova.network.neutron [req-49868658-dba1-4318-a820-c2f145163118 req-fcab777a-6479-4ebd-b52e-0bcc45cb7f8b service nova] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Refreshing network info cache for port f6501e07-0e63-4597-86b9-bbe243fabaa3 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 913.534272] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:94:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6501e07-0e63-4597-86b9-bbe243fabaa3', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.544953] env[63293]: DEBUG oslo.service.loopingcall [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.545542] env[63293]: DEBUG nova.network.neutron [-] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.548222] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 913.549075] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-01943861-7aac-4337-90cb-ac67f27115eb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.570262] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-496c4305-bf39-4ed2-be8b-d91b6ebf8f53 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.580077] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7dbee7-a402-487a-9f98-a020078e526a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.592913] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.592913] env[63293]: value = "task-1327864" [ 913.592913] env[63293]: _type = "Task" [ 913.592913] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.602647] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327864, 'name': CreateVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.610287] env[63293]: DEBUG nova.compute.manager [req-93268fa7-d256-4bbc-bbda-ae66c9def0e7 req-960dbad3-401b-41fa-9df1-6782f0c9404d service nova] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Detach interface failed, port_id=7918ba6d-5544-486f-a409-ab8077d44efc, reason: Instance 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 913.700694] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327859, 'name': ReconfigVM_Task, 'duration_secs': 0.178683} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.700876] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Reconfigured VM instance instance-0000004e to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 913.701616] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f046717-7898-4da2-9a99-c3d7d811c053 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.725139] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 76ebe2ec-332c-4d3a-957e-d0c928dafdc7/76ebe2ec-332c-4d3a-957e-d0c928dafdc7.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.726425] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68934b58-1271-4e03-a562-1693c398f9c0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.747460] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 913.747460] env[63293]: value = "task-1327865" [ 913.747460] env[63293]: _type = "Task" [ 913.747460] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.757349] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327865, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.022213] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquiring lock "refresh_cache-8625a358-ec48-46b9-8738-c49a2ba58362" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.022393] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquired lock "refresh_cache-8625a358-ec48-46b9-8738-c49a2ba58362" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.022558] env[63293]: DEBUG nova.network.neutron [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.049716] env[63293]: INFO nova.compute.manager [-] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Took 1.30 seconds to deallocate network for instance. [ 914.074290] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff4337eb-0753-412b-9b98-399b9d58bea7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.081948] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8bab3d-0078-4fdc-a9df-427f58853460 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.119414] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac303c47-770d-48e8-839b-6fe22b7eae98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.130709] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637d235e-7ad0-4d2d-a712-073e2acc4fb0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.134465] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327864, 'name': CreateVM_Task, 'duration_secs': 0.336065} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.135238] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 914.136713] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.136713] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.137162] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.145078] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cb19166-8def-47fd-a353-9c170df6280e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.147221] env[63293]: DEBUG nova.compute.provider_tree [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.153918] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 914.153918] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523431f4-9abe-aaa3-30bb-15f02e4afe16" [ 914.153918] env[63293]: _type = "Task" [ 914.153918] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.162927] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523431f4-9abe-aaa3-30bb-15f02e4afe16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.243077] env[63293]: DEBUG nova.objects.instance [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'flavor' on Instance uuid 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.257383] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327865, 'name': ReconfigVM_Task, 'duration_secs': 0.364315} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.257618] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 76ebe2ec-332c-4d3a-957e-d0c928dafdc7/76ebe2ec-332c-4d3a-957e-d0c928dafdc7.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.257905] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance '76ebe2ec-332c-4d3a-957e-d0c928dafdc7' progress to 50 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 914.287258] env[63293]: DEBUG nova.network.neutron [req-49868658-dba1-4318-a820-c2f145163118 req-fcab777a-6479-4ebd-b52e-0bcc45cb7f8b service nova] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Updated VIF entry in instance network info cache for port f6501e07-0e63-4597-86b9-bbe243fabaa3. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 914.287679] env[63293]: DEBUG nova.network.neutron [req-49868658-dba1-4318-a820-c2f145163118 req-fcab777a-6479-4ebd-b52e-0bcc45cb7f8b service nova] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Updating instance_info_cache with network_info: [{"id": "f6501e07-0e63-4597-86b9-bbe243fabaa3", "address": "fa:16:3e:06:94:5f", "network": {"id": "02c216f2-c41a-4722-91df-330cd520be3d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f57886847a1446109c6ab061a248dcf0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6501e07-0e", "ovs_interfaceid": "f6501e07-0e63-4597-86b9-bbe243fabaa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.557734] env[63293]: DEBUG nova.network.neutron [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 914.560443] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.653060] env[63293]: DEBUG nova.scheduler.client.report [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.674473] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523431f4-9abe-aaa3-30bb-15f02e4afe16, 'name': SearchDatastore_Task, 'duration_secs': 0.010017} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.675506] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.675755] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.676249] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.676369] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.676727] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.677258] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5561dcf8-a923-452e-891d-f1705115c005 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.686973] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.687266] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.688028] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71d15f79-a7df-4b94-895d-a1dc6e2c9618 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.696491] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 914.696491] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52633b44-37ae-7f3f-03f8-ead6d8b03825" [ 914.696491] env[63293]: _type = "Task" [ 914.696491] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.706055] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52633b44-37ae-7f3f-03f8-ead6d8b03825, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.750145] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.750357] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquired lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.750585] env[63293]: DEBUG nova.network.neutron [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.750789] env[63293]: DEBUG nova.objects.instance [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'info_cache' on Instance uuid 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.764659] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c7a325-35ca-443d-b236-f369f1d97803 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.789510] env[63293]: DEBUG nova.network.neutron [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Updating instance_info_cache with network_info: [{"id": "9551ebb5-552d-47d0-a436-994fe2b1d00b", "address": "fa:16:3e:96:b9:15", "network": {"id": "922225f5-dfb0-4ae5-ad42-460a6978dd02", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-380130906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2bb8a8bec2a04cb292af72d1216bb355", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9551ebb5-55", "ovs_interfaceid": "9551ebb5-552d-47d0-a436-994fe2b1d00b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.792823] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f500392-9c96-49d0-bab6-56d9aa041945 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.797756] env[63293]: DEBUG oslo_concurrency.lockutils [req-49868658-dba1-4318-a820-c2f145163118 req-fcab777a-6479-4ebd-b52e-0bcc45cb7f8b service nova] Releasing lock "refresh_cache-5b1264b4-2554-4c62-9ff9-4616b03d3609" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.814043] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance '76ebe2ec-332c-4d3a-957e-d0c928dafdc7' progress to 67 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 915.084437] env[63293]: DEBUG nova.compute.manager [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Received event network-vif-plugged-9551ebb5-552d-47d0-a436-994fe2b1d00b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.084437] env[63293]: DEBUG oslo_concurrency.lockutils [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] Acquiring lock "8625a358-ec48-46b9-8738-c49a2ba58362-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.084437] env[63293]: DEBUG oslo_concurrency.lockutils [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] Lock "8625a358-ec48-46b9-8738-c49a2ba58362-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.084437] env[63293]: DEBUG oslo_concurrency.lockutils [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] Lock "8625a358-ec48-46b9-8738-c49a2ba58362-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.084437] env[63293]: DEBUG nova.compute.manager [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] No waiting events found dispatching network-vif-plugged-9551ebb5-552d-47d0-a436-994fe2b1d00b {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 915.084437] env[63293]: WARNING nova.compute.manager [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Received unexpected event network-vif-plugged-9551ebb5-552d-47d0-a436-994fe2b1d00b for instance with vm_state building and task_state spawning. [ 915.084437] env[63293]: DEBUG nova.compute.manager [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Received event network-changed-9551ebb5-552d-47d0-a436-994fe2b1d00b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.084437] env[63293]: DEBUG nova.compute.manager [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Refreshing instance network info cache due to event network-changed-9551ebb5-552d-47d0-a436-994fe2b1d00b. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 915.084437] env[63293]: DEBUG oslo_concurrency.lockutils [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] Acquiring lock "refresh_cache-8625a358-ec48-46b9-8738-c49a2ba58362" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.160087] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.160307] env[63293]: DEBUG nova.compute.manager [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.162821] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.191s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.164275] env[63293]: INFO nova.compute.claims [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.210539] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52633b44-37ae-7f3f-03f8-ead6d8b03825, 'name': SearchDatastore_Task, 'duration_secs': 0.00947} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.211598] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3abe4dbe-e867-4a45-bab5-02fd3c9ede58 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.217727] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 915.217727] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fdcb67-20a1-375c-4ef6-6e27ebce20f4" [ 915.217727] env[63293]: _type = "Task" [ 915.217727] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.226395] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fdcb67-20a1-375c-4ef6-6e27ebce20f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.253986] env[63293]: DEBUG nova.objects.base [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Object Instance<0977c79a-2ef3-4ec9-a0a0-de1ed5799723> lazy-loaded attributes: flavor,info_cache {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 915.298785] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Releasing lock "refresh_cache-8625a358-ec48-46b9-8738-c49a2ba58362" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.299127] env[63293]: DEBUG nova.compute.manager [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Instance network_info: |[{"id": "9551ebb5-552d-47d0-a436-994fe2b1d00b", "address": "fa:16:3e:96:b9:15", "network": {"id": "922225f5-dfb0-4ae5-ad42-460a6978dd02", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-380130906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2bb8a8bec2a04cb292af72d1216bb355", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9551ebb5-55", "ovs_interfaceid": "9551ebb5-552d-47d0-a436-994fe2b1d00b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 915.299464] env[63293]: DEBUG oslo_concurrency.lockutils [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] Acquired lock "refresh_cache-8625a358-ec48-46b9-8738-c49a2ba58362" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.299730] env[63293]: DEBUG nova.network.neutron [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Refreshing network info cache for port 9551ebb5-552d-47d0-a436-994fe2b1d00b {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 915.301677] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:b9:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e41070eb-3ac1-4ca9-a3d0-fd65893a97de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9551ebb5-552d-47d0-a436-994fe2b1d00b', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.309341] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Creating folder: Project (2bb8a8bec2a04cb292af72d1216bb355). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 915.310121] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-97e99235-7bf8-4d45-9c60-f2d4e5f9ea78 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.323390] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Created folder: Project (2bb8a8bec2a04cb292af72d1216bb355) in parent group-v283678. [ 915.323621] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Creating folder: Instances. Parent ref: group-v283795. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 915.324200] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aee01a93-04cc-4e29-bb60-0830838f03b5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.335768] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Created folder: Instances in parent group-v283795. [ 915.336033] env[63293]: DEBUG oslo.service.loopingcall [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.336233] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 915.336453] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91b642c1-a27f-4319-ac28-3290e230cf09 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.356033] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.356033] env[63293]: value = "task-1327868" [ 915.356033] env[63293]: _type = "Task" [ 915.356033] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.356859] env[63293]: DEBUG nova.network.neutron [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Port 46d5563f-41a2-48ab-9d32-3aa394a13da8 binding to destination host cpu-1 is already ACTIVE {{(pid=63293) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 915.365886] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327868, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.653445] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Volume attach. Driver type: vmdk {{(pid=63293) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 915.653756] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283791', 'volume_id': '6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'name': 'volume-6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd753598d-a92a-4515-9ad1-d386294f7a99', 'attached_at': '', 'detached_at': '', 'volume_id': '6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'serial': '6c9dda44-62d5-4ee9-8961-4b9fec04802a'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 915.654779] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef8bcda-a0b8-4d54-a77c-bfadf5a83616 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.674546] env[63293]: DEBUG nova.compute.utils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.678599] env[63293]: DEBUG nova.compute.manager [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.682162] env[63293]: DEBUG nova.network.neutron [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 915.682162] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d721cbd-1c25-474e-9396-082630b2034d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.714443] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] volume-6c9dda44-62d5-4ee9-8961-4b9fec04802a/volume-6c9dda44-62d5-4ee9-8961-4b9fec04802a.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.715237] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9d08d94-6935-4128-a596-ccc9203bf610 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.741286] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fdcb67-20a1-375c-4ef6-6e27ebce20f4, 'name': SearchDatastore_Task, 'duration_secs': 0.014991} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.743804] env[63293]: DEBUG nova.policy [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd2a9495ab0b4768a4075a71f25aa9a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff8479fc251544e9b4618deccf992754', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.745402] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.745696] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 5b1264b4-2554-4c62-9ff9-4616b03d3609/5b1264b4-2554-4c62-9ff9-4616b03d3609.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 915.746111] env[63293]: DEBUG oslo_vmware.api [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 915.746111] env[63293]: value = "task-1327870" [ 915.746111] env[63293]: _type = "Task" [ 915.746111] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.746319] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a3b93a12-10ac-466a-8eb8-124f953f3a8e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.760606] env[63293]: DEBUG oslo_vmware.api [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327870, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.762138] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 915.762138] env[63293]: value = "task-1327871" [ 915.762138] env[63293]: _type = "Task" [ 915.762138] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.772371] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327871, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.871764] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327868, 'name': CreateVM_Task, 'duration_secs': 0.385859} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.875020] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.875020] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.875020] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.875020] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.875020] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f445558e-6ac8-49cc-baf0-b63f0c6348dd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.879027] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for the task: (returnval){ [ 915.879027] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52166881-46cc-861d-46c0-851a68baabad" [ 915.879027] env[63293]: _type = "Task" [ 915.879027] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.892468] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52166881-46cc-861d-46c0-851a68baabad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.052181] env[63293]: DEBUG nova.network.neutron [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Updated VIF entry in instance network info cache for port 9551ebb5-552d-47d0-a436-994fe2b1d00b. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 916.052181] env[63293]: DEBUG nova.network.neutron [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Updating instance_info_cache with network_info: [{"id": "9551ebb5-552d-47d0-a436-994fe2b1d00b", "address": "fa:16:3e:96:b9:15", "network": {"id": "922225f5-dfb0-4ae5-ad42-460a6978dd02", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-380130906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2bb8a8bec2a04cb292af72d1216bb355", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9551ebb5-55", "ovs_interfaceid": "9551ebb5-552d-47d0-a436-994fe2b1d00b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.182074] env[63293]: DEBUG nova.compute.manager [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.187227] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "35cbf277-b85a-4101-beae-0125fb5a61f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.187802] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "35cbf277-b85a-4101-beae-0125fb5a61f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.233937] env[63293]: DEBUG nova.network.neutron [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updating instance_info_cache with network_info: [{"id": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "address": "fa:16:3e:80:11:9c", "network": {"id": "9d5a3e58-567a-4c3f-9ad9-9fcbf81fced9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-949261114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1a30d37c1b47fd8670db0a9ef7020d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f0bfb1c-61", "ovs_interfaceid": "2f0bfb1c-6184-49f1-9c71-cb7459e1470d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.264406] env[63293]: DEBUG oslo_vmware.api [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327870, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.280095] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327871, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.294144] env[63293]: DEBUG nova.network.neutron [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Successfully created port: 94365438-f2bc-4062-add0-5c8519bf1a6d {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.392070] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.392326] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.392507] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.401669] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52166881-46cc-861d-46c0-851a68baabad, 'name': SearchDatastore_Task, 'duration_secs': 0.014719} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.402042] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.402483] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 916.402734] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.402930] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.403085] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.403354] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0e153d48-4b49-4724-a7bc-ca4dd48aa87f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.412528] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.412811] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 916.414214] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8deff779-199d-4bd1-8133-3a7eb4aa846c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.424424] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for the task: (returnval){ [ 916.424424] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5237e139-aa77-553d-291a-d749a4c9ed2e" [ 916.424424] env[63293]: _type = "Task" [ 916.424424] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.433157] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5237e139-aa77-553d-291a-d749a4c9ed2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.499805] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df880b3-877a-4a9e-89c9-d89254d41271 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.508341] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5a92ed-30c4-4aec-8225-236a9f76f832 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.540660] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc6fb85-c19d-46a6-acd1-b4ea41a7efec {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.548505] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14031bb5-baac-4836-afa3-dcc9b5d38d4f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.554669] env[63293]: DEBUG oslo_concurrency.lockutils [req-e929d8e5-06ef-4d0f-bb00-8d547afc619e req-9a31ac0c-67ea-4508-8060-1d63006753c9 service nova] Releasing lock "refresh_cache-8625a358-ec48-46b9-8738-c49a2ba58362" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.563510] env[63293]: DEBUG nova.compute.provider_tree [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.693935] env[63293]: DEBUG nova.compute.manager [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 916.735017] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Releasing lock "refresh_cache-0977c79a-2ef3-4ec9-a0a0-de1ed5799723" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.759918] env[63293]: DEBUG oslo_vmware.api [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327870, 'name': ReconfigVM_Task, 'duration_secs': 0.620417} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.760227] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Reconfigured VM instance instance-0000004a to attach disk [datastore1] volume-6c9dda44-62d5-4ee9-8961-4b9fec04802a/volume-6c9dda44-62d5-4ee9-8961-4b9fec04802a.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.766511] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c530e99d-b158-4aca-a420-5b52d04c697a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.785259] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327871, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.62221} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.786615] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 5b1264b4-2554-4c62-9ff9-4616b03d3609/5b1264b4-2554-4c62-9ff9-4616b03d3609.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 916.786851] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.787183] env[63293]: DEBUG oslo_vmware.api [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 916.787183] env[63293]: value = "task-1327872" [ 916.787183] env[63293]: _type = "Task" [ 916.787183] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.787456] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-744f3eff-96bf-4c1b-b197-53331e253a8e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.796925] env[63293]: DEBUG oslo_vmware.api [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327872, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.798138] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 916.798138] env[63293]: value = "task-1327873" [ 916.798138] env[63293]: _type = "Task" [ 916.798138] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.805399] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327873, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.936337] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5237e139-aa77-553d-291a-d749a4c9ed2e, 'name': SearchDatastore_Task, 'duration_secs': 0.009732} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.937764] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9eaec0d1-438d-4227-ab0c-63bf24410a9b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.943698] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for the task: (returnval){ [ 916.943698] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528c5250-2e14-cc17-70ae-414cfcc370b3" [ 916.943698] env[63293]: _type = "Task" [ 916.943698] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.952138] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528c5250-2e14-cc17-70ae-414cfcc370b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.066767] env[63293]: DEBUG nova.scheduler.client.report [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.193012] env[63293]: DEBUG nova.compute.manager [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.212587] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.216398] env[63293]: DEBUG nova.virt.hardware [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.216398] env[63293]: DEBUG nova.virt.hardware [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.216585] env[63293]: DEBUG nova.virt.hardware [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.217445] env[63293]: DEBUG nova.virt.hardware [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.217445] env[63293]: DEBUG nova.virt.hardware [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.217445] env[63293]: DEBUG nova.virt.hardware [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.217445] env[63293]: DEBUG nova.virt.hardware [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.217445] env[63293]: DEBUG nova.virt.hardware [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.217804] env[63293]: DEBUG nova.virt.hardware [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.218013] env[63293]: DEBUG nova.virt.hardware [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.218229] env[63293]: DEBUG nova.virt.hardware [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.219511] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916b529e-64ec-45b0-99c3-bbddf59e5214 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.228950] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f08e49-254f-4104-9ebc-152c463a302a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.244014] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.244549] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72584c51-fec8-47bb-bce6-1e76cb5adaf9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.251204] env[63293]: DEBUG oslo_vmware.api [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 917.251204] env[63293]: value = "task-1327874" [ 917.251204] env[63293]: _type = "Task" [ 917.251204] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.259434] env[63293]: DEBUG oslo_vmware.api [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327874, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.299402] env[63293]: DEBUG oslo_vmware.api [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327872, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.307495] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327873, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097918} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.307774] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.308594] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c78ed9-ce3c-4ee8-9329-350d58dff7ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.330093] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 5b1264b4-2554-4c62-9ff9-4616b03d3609/5b1264b4-2554-4c62-9ff9-4616b03d3609.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.330381] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-883f8477-9102-4edb-935d-2998dd07f36e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.351556] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 917.351556] env[63293]: value = "task-1327875" [ 917.351556] env[63293]: _type = "Task" [ 917.351556] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.361966] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327875, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.454518] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528c5250-2e14-cc17-70ae-414cfcc370b3, 'name': SearchDatastore_Task, 'duration_secs': 0.009442} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.454750] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.455030] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 8625a358-ec48-46b9-8738-c49a2ba58362/8625a358-ec48-46b9-8738-c49a2ba58362.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 917.455312] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90e897cf-1a8c-47f5-8603-176712da7afe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.462713] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for the task: (returnval){ [ 917.462713] env[63293]: value = "task-1327876" [ 917.462713] env[63293]: _type = "Task" [ 917.462713] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.470829] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327876, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.474852] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.475038] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.475216] env[63293]: DEBUG nova.network.neutron [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.573505] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.574247] env[63293]: DEBUG nova.compute.manager [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.577423] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.368s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.579519] env[63293]: INFO nova.compute.claims [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.768724] env[63293]: DEBUG oslo_vmware.api [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327874, 'name': PowerOnVM_Task, 'duration_secs': 0.430171} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.769057] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 917.769277] env[63293]: DEBUG nova.compute.manager [None req-ef7fd088-b7ad-4ccc-b060-dd8d0b3f7089 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.770182] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b71072-4eee-4971-a23a-337e588751f2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.802129] env[63293]: DEBUG oslo_vmware.api [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327872, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.863152] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327875, 'name': ReconfigVM_Task, 'duration_secs': 0.331784} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.864815] env[63293]: DEBUG nova.compute.manager [req-2e2a9146-3499-4108-9349-3ca3f2e6b211 req-008a314e-fd69-409a-9cc1-1c304bba3655 service nova] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Received event network-vif-plugged-94365438-f2bc-4062-add0-5c8519bf1a6d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.865111] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e2a9146-3499-4108-9349-3ca3f2e6b211 req-008a314e-fd69-409a-9cc1-1c304bba3655 service nova] Acquiring lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.865504] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e2a9146-3499-4108-9349-3ca3f2e6b211 req-008a314e-fd69-409a-9cc1-1c304bba3655 service nova] Lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.865698] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e2a9146-3499-4108-9349-3ca3f2e6b211 req-008a314e-fd69-409a-9cc1-1c304bba3655 service nova] Lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.866061] env[63293]: DEBUG nova.compute.manager [req-2e2a9146-3499-4108-9349-3ca3f2e6b211 req-008a314e-fd69-409a-9cc1-1c304bba3655 service nova] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] No waiting events found dispatching network-vif-plugged-94365438-f2bc-4062-add0-5c8519bf1a6d {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 917.866323] env[63293]: WARNING nova.compute.manager [req-2e2a9146-3499-4108-9349-3ca3f2e6b211 req-008a314e-fd69-409a-9cc1-1c304bba3655 service nova] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Received unexpected event network-vif-plugged-94365438-f2bc-4062-add0-5c8519bf1a6d for instance with vm_state building and task_state spawning. [ 917.866795] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 5b1264b4-2554-4c62-9ff9-4616b03d3609/5b1264b4-2554-4c62-9ff9-4616b03d3609.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.868037] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6e6b065-2071-4f63-8ddb-6935424e3589 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.876601] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 917.876601] env[63293]: value = "task-1327877" [ 917.876601] env[63293]: _type = "Task" [ 917.876601] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.887019] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327877, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.974134] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327876, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.013283] env[63293]: DEBUG nova.network.neutron [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Successfully updated port: 94365438-f2bc-4062-add0-5c8519bf1a6d {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.084200] env[63293]: DEBUG nova.compute.utils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.089758] env[63293]: DEBUG nova.compute.manager [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 918.089937] env[63293]: DEBUG nova.network.neutron [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 918.170770] env[63293]: DEBUG nova.policy [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc557354560049119144c62dd61d0816', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2af826d819b4cc4ab27ff8b95144039', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 918.300763] env[63293]: DEBUG oslo_vmware.api [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327872, 'name': ReconfigVM_Task, 'duration_secs': 1.189567} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.302467] env[63293]: DEBUG nova.network.neutron [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance_info_cache with network_info: [{"id": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "address": "fa:16:3e:12:25:eb", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46d5563f-41", "ovs_interfaceid": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.306035] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283791', 'volume_id': '6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'name': 'volume-6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd753598d-a92a-4515-9ad1-d386294f7a99', 'attached_at': '', 'detached_at': '', 'volume_id': '6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'serial': '6c9dda44-62d5-4ee9-8961-4b9fec04802a'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 918.391681] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327877, 'name': Rename_Task, 'duration_secs': 0.260219} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.392058] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 918.392379] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96f4fb1f-4d54-4216-b24a-eee496cce84f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.400016] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 918.400016] env[63293]: value = "task-1327878" [ 918.400016] env[63293]: _type = "Task" [ 918.400016] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.409163] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327878, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.474890] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327876, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.559319} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.476980] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 8625a358-ec48-46b9-8738-c49a2ba58362/8625a358-ec48-46b9-8738-c49a2ba58362.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 918.476980] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.476980] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70f2347f-7660-4454-9c7b-82de1dda474b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.485066] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for the task: (returnval){ [ 918.485066] env[63293]: value = "task-1327879" [ 918.485066] env[63293]: _type = "Task" [ 918.485066] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.493970] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327879, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.516396] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-fe55a9b7-28b5-4b04-84f1-6b85b579a144" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.516396] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-fe55a9b7-28b5-4b04-84f1-6b85b579a144" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.516396] env[63293]: DEBUG nova.network.neutron [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 918.553803] env[63293]: DEBUG nova.network.neutron [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Successfully created port: e54443d3-84a6-4e40-8220-bb2079d7805c {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.590348] env[63293]: DEBUG nova.compute.manager [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.689874] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.791663] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52017786-2c19-5c20-2ae3-52b86b0c3804/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 918.799931] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9c977b-8313-491d-811c-935a18e092d4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.804418] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52017786-2c19-5c20-2ae3-52b86b0c3804/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 918.804599] env[63293]: ERROR oslo_vmware.rw_handles [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52017786-2c19-5c20-2ae3-52b86b0c3804/disk-0.vmdk due to incomplete transfer. [ 918.804838] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0508def6-882c-4451-918e-ecbe679fca0d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.806639] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.822374] env[63293]: DEBUG oslo_vmware.rw_handles [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52017786-2c19-5c20-2ae3-52b86b0c3804/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 918.822616] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Uploaded image 309b9adc-458f-4eaa-8a37-2f7ffc94d5dc to the Glance image server {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 918.824807] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Destroying the VM {{(pid=63293) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 918.825108] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-acc58ca0-7b70-4f6c-9088-dfd8c2fed143 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.834787] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 918.834787] env[63293]: value = "task-1327880" [ 918.834787] env[63293]: _type = "Task" [ 918.834787] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.851442] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327880, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.895054] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6c2d7d-45f7-4f30-b7eb-2e346a41c8fb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.907445] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2574ea3-2696-4e84-99e1-2f4677863510 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.913727] env[63293]: DEBUG oslo_vmware.api [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327878, 'name': PowerOnVM_Task, 'duration_secs': 0.500624} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.914401] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 918.914614] env[63293]: INFO nova.compute.manager [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Took 8.38 seconds to spawn the instance on the hypervisor. [ 918.915253] env[63293]: DEBUG nova.compute.manager [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.915858] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91cf527e-372c-4992-9fc3-7fc785e2abe3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.943755] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bb7a96-6fa6-4c01-9c01-b02a0f70e885 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.955387] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2747d3-c29c-4f38-b8b7-50ff29c99f5b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.971033] env[63293]: DEBUG nova.compute.provider_tree [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.994185] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327879, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071116} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.994519] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.995320] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d865ff47-7d12-440a-b40e-4c5b3b18338e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.016786] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 8625a358-ec48-46b9-8738-c49a2ba58362/8625a358-ec48-46b9-8738-c49a2ba58362.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.017099] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a18bd5af-a0df-4d0a-a396-41ab97dc19ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.038012] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for the task: (returnval){ [ 919.038012] env[63293]: value = "task-1327881" [ 919.038012] env[63293]: _type = "Task" [ 919.038012] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.045861] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327881, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.061214] env[63293]: DEBUG nova.network.neutron [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 919.334150] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ea4ae9-e133-465c-b11d-9e5a22736fb2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.361744] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce01032-59dd-47f4-8e5f-d96ffe896e3a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.364511] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327880, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.369064] env[63293]: DEBUG nova.objects.instance [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lazy-loading 'flavor' on Instance uuid d753598d-a92a-4515-9ad1-d386294f7a99 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.370544] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance '76ebe2ec-332c-4d3a-957e-d0c928dafdc7' progress to 83 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 919.457930] env[63293]: INFO nova.compute.manager [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Took 29.19 seconds to build instance. [ 919.474369] env[63293]: DEBUG nova.scheduler.client.report [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.483017] env[63293]: DEBUG nova.network.neutron [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Updating instance_info_cache with network_info: [{"id": "94365438-f2bc-4062-add0-5c8519bf1a6d", "address": "fa:16:3e:4a:8a:42", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94365438-f2", "ovs_interfaceid": "94365438-f2bc-4062-add0-5c8519bf1a6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.555558] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327881, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.607367] env[63293]: DEBUG nova.compute.manager [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.628162] env[63293]: DEBUG nova.virt.hardware [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='cfa393e0151a51d6470258c6f9706ca7',container_format='bare',created_at=2024-10-21T17:39:48Z,direct_url=,disk_format='vmdk',id=8b176c14-05dc-43c1-a475-57283dc96350,min_disk=1,min_ram=0,name='tempest-test-snap-1489793634',owner='a2af826d819b4cc4ab27ff8b95144039',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-21T17:40:03Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.628162] env[63293]: DEBUG nova.virt.hardware [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.628162] env[63293]: DEBUG nova.virt.hardware [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.628162] env[63293]: DEBUG nova.virt.hardware [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.628162] env[63293]: DEBUG nova.virt.hardware [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.628162] env[63293]: DEBUG nova.virt.hardware [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.628162] env[63293]: DEBUG nova.virt.hardware [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.628753] env[63293]: DEBUG nova.virt.hardware [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.629113] env[63293]: DEBUG nova.virt.hardware [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.629510] env[63293]: DEBUG nova.virt.hardware [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.629874] env[63293]: DEBUG nova.virt.hardware [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.631361] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d726cc-bc60-46d1-b737-e0fac90f4af8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.640978] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20322697-dae6-4ffa-b39f-714d4dcb268e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.815677] env[63293]: DEBUG nova.network.neutron [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Successfully updated port: e54443d3-84a6-4e40-8220-bb2079d7805c {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 919.848999] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327880, 'name': Destroy_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.878991] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.879690] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4d9ff88e-222e-4b92-9529-f91cfcb4032a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.891s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.880843] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69b97c91-29a6-4127-95e4-549f91609e58 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.887497] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 919.887497] env[63293]: value = "task-1327882" [ 919.887497] env[63293]: _type = "Task" [ 919.887497] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.895122] env[63293]: DEBUG nova.compute.manager [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Received event network-changed-94365438-f2bc-4062-add0-5c8519bf1a6d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.895122] env[63293]: DEBUG nova.compute.manager [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Refreshing instance network info cache due to event network-changed-94365438-f2bc-4062-add0-5c8519bf1a6d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 919.895122] env[63293]: DEBUG oslo_concurrency.lockutils [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] Acquiring lock "refresh_cache-fe55a9b7-28b5-4b04-84f1-6b85b579a144" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.898775] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327882, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.960477] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c6766571-3774-4106-b5a2-f6d76ca66bf5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "5b1264b4-2554-4c62-9ff9-4616b03d3609" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.699s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.980865] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.985410] env[63293]: DEBUG nova.compute.manager [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 919.986988] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-fe55a9b7-28b5-4b04-84f1-6b85b579a144" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.987341] env[63293]: DEBUG nova.compute.manager [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Instance network_info: |[{"id": "94365438-f2bc-4062-add0-5c8519bf1a6d", "address": "fa:16:3e:4a:8a:42", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94365438-f2", "ovs_interfaceid": "94365438-f2bc-4062-add0-5c8519bf1a6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 919.987676] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.427s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.987906] env[63293]: DEBUG nova.objects.instance [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lazy-loading 'resources' on Instance uuid 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.988900] env[63293]: DEBUG oslo_concurrency.lockutils [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] Acquired lock "refresh_cache-fe55a9b7-28b5-4b04-84f1-6b85b579a144" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.989179] env[63293]: DEBUG nova.network.neutron [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Refreshing network info cache for port 94365438-f2bc-4062-add0-5c8519bf1a6d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 919.990169] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:8a:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94365438-f2bc-4062-add0-5c8519bf1a6d', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.997679] env[63293]: DEBUG oslo.service.loopingcall [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.998852] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 919.999172] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c85fe6b-b4ee-4725-87c6-4dd21ae80d1a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.019191] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.019191] env[63293]: value = "task-1327883" [ 920.019191] env[63293]: _type = "Task" [ 920.019191] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.027795] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327883, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.051439] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327881, 'name': ReconfigVM_Task, 'duration_secs': 0.87047} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.051726] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 8625a358-ec48-46b9-8738-c49a2ba58362/8625a358-ec48-46b9-8738-c49a2ba58362.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.052390] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-53d0ab59-4154-4a19-9480-43a39d3220a3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.058829] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for the task: (returnval){ [ 920.058829] env[63293]: value = "task-1327884" [ 920.058829] env[63293]: _type = "Task" [ 920.058829] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.069309] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327884, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.318393] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "refresh_cache-2557bded-6ccb-4c0f-8b97-34529b4bb740" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.318506] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "refresh_cache-2557bded-6ccb-4c0f-8b97-34529b4bb740" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.320106] env[63293]: DEBUG nova.network.neutron [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 920.346842] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327880, 'name': Destroy_Task, 'duration_secs': 1.041849} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.347857] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Destroyed the VM [ 920.348250] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Deleting Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 920.348708] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0f0df1ef-22d5-49aa-b64d-ddde1423c8ce {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.355740] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 920.355740] env[63293]: value = "task-1327885" [ 920.355740] env[63293]: _type = "Task" [ 920.355740] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.365168] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327885, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.397631] env[63293]: DEBUG oslo_vmware.api [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327882, 'name': PowerOnVM_Task, 'duration_secs': 0.447245} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.398017] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.398106] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed7b00a-2d08-49e1-9fd1-b983a5ac80f0 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance '76ebe2ec-332c-4d3a-957e-d0c928dafdc7' progress to 100 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 920.505229] env[63293]: DEBUG nova.compute.utils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.506346] env[63293]: DEBUG nova.compute.manager [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 920.506498] env[63293]: DEBUG nova.network.neutron [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 920.532033] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327883, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.568274] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327884, 'name': Rename_Task, 'duration_secs': 0.16759} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.571202] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.573934] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d74b8998-eef7-4231-82ba-1df60b316987 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.580034] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for the task: (returnval){ [ 920.580034] env[63293]: value = "task-1327886" [ 920.580034] env[63293]: _type = "Task" [ 920.580034] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.588781] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327886, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.608756] env[63293]: DEBUG nova.policy [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd40ed85d59b1470481ca4dff87e4832c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eec85c5992d64d62a90e290a1aa2d441', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 920.670905] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d1c855-8c50-43db-91dd-5d46a878da20 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.683315] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fb756369-03e8-43bf-b6da-591c456d0130 tempest-ServersAdminNegativeTestJSON-458761815 tempest-ServersAdminNegativeTestJSON-458761815-project-admin] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Suspending the VM {{(pid=63293) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 920.683669] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-bd9cde4f-4082-4fc8-9602-fb5bd0793055 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.691174] env[63293]: DEBUG oslo_vmware.api [None req-fb756369-03e8-43bf-b6da-591c456d0130 tempest-ServersAdminNegativeTestJSON-458761815 tempest-ServersAdminNegativeTestJSON-458761815-project-admin] Waiting for the task: (returnval){ [ 920.691174] env[63293]: value = "task-1327887" [ 920.691174] env[63293]: _type = "Task" [ 920.691174] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.700547] env[63293]: DEBUG oslo_vmware.api [None req-fb756369-03e8-43bf-b6da-591c456d0130 tempest-ServersAdminNegativeTestJSON-458761815 tempest-ServersAdminNegativeTestJSON-458761815-project-admin] Task: {'id': task-1327887, 'name': SuspendVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.768942] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ff4ee8-e149-4ba1-a00e-6bd933be0f66 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.780998] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e156a0ad-92b4-41de-ac2b-551921a5595e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.818761] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5deaa9-a763-4cb3-87f5-93ffa8add618 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.827837] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59742c0-e714-4bbc-a05b-0300841c11b6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.842115] env[63293]: DEBUG nova.compute.provider_tree [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.866835] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327885, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.883366] env[63293]: DEBUG nova.network.neutron [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 921.009848] env[63293]: DEBUG nova.compute.manager [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 921.032560] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327883, 'name': CreateVM_Task, 'duration_secs': 0.960129} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.032771] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 921.033514] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.033854] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.034240] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.034520] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5165a44b-813e-4438-a04a-d875fd4f466b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.039530] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 921.039530] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522fe73a-c24e-aeca-46fa-1e26bf55f953" [ 921.039530] env[63293]: _type = "Task" [ 921.039530] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.050433] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522fe73a-c24e-aeca-46fa-1e26bf55f953, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.085952] env[63293]: DEBUG nova.network.neutron [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Updated VIF entry in instance network info cache for port 94365438-f2bc-4062-add0-5c8519bf1a6d. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 921.086331] env[63293]: DEBUG nova.network.neutron [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Updating instance_info_cache with network_info: [{"id": "94365438-f2bc-4062-add0-5c8519bf1a6d", "address": "fa:16:3e:4a:8a:42", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94365438-f2", "ovs_interfaceid": "94365438-f2bc-4062-add0-5c8519bf1a6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.093855] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327886, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.094781] env[63293]: DEBUG nova.network.neutron [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Updating instance_info_cache with network_info: [{"id": "e54443d3-84a6-4e40-8220-bb2079d7805c", "address": "fa:16:3e:d3:8e:0c", "network": {"id": "340bfd3b-7dff-454b-87df-788003988adc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-388838680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2af826d819b4cc4ab27ff8b95144039", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape54443d3-84", "ovs_interfaceid": "e54443d3-84a6-4e40-8220-bb2079d7805c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.203395] env[63293]: DEBUG oslo_vmware.api [None req-fb756369-03e8-43bf-b6da-591c456d0130 tempest-ServersAdminNegativeTestJSON-458761815 tempest-ServersAdminNegativeTestJSON-458761815-project-admin] Task: {'id': task-1327887, 'name': SuspendVM_Task} progress is 54%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.281041] env[63293]: DEBUG nova.network.neutron [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Successfully created port: 4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 921.347929] env[63293]: DEBUG nova.scheduler.client.report [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.368260] env[63293]: DEBUG oslo_vmware.api [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327885, 'name': RemoveSnapshot_Task, 'duration_secs': 0.899883} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.368552] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Deleted Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 921.368793] env[63293]: INFO nova.compute.manager [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Took 14.74 seconds to snapshot the instance on the hypervisor. [ 921.552918] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522fe73a-c24e-aeca-46fa-1e26bf55f953, 'name': SearchDatastore_Task, 'duration_secs': 0.051288} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.553285] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.553780] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.554066] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.554232] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.554421] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.554699] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f1d6ce7-e219-47a5-8d63-3da90b2ca75b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.567718] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.567910] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.568718] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b068a41-33b6-4b96-ba63-310f39191406 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.575312] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 921.575312] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d4dcdb-4c38-56ee-a036-37a877f10120" [ 921.575312] env[63293]: _type = "Task" [ 921.575312] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.588735] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d4dcdb-4c38-56ee-a036-37a877f10120, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.589631] env[63293]: DEBUG oslo_concurrency.lockutils [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] Releasing lock "refresh_cache-fe55a9b7-28b5-4b04-84f1-6b85b579a144" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.589969] env[63293]: DEBUG nova.compute.manager [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Received event network-vif-plugged-e54443d3-84a6-4e40-8220-bb2079d7805c {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.590286] env[63293]: DEBUG oslo_concurrency.lockutils [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] Acquiring lock "2557bded-6ccb-4c0f-8b97-34529b4bb740-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.590602] env[63293]: DEBUG oslo_concurrency.lockutils [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] Lock "2557bded-6ccb-4c0f-8b97-34529b4bb740-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.590886] env[63293]: DEBUG oslo_concurrency.lockutils [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] Lock "2557bded-6ccb-4c0f-8b97-34529b4bb740-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.591103] env[63293]: DEBUG nova.compute.manager [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] No waiting events found dispatching network-vif-plugged-e54443d3-84a6-4e40-8220-bb2079d7805c {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.591288] env[63293]: WARNING nova.compute.manager [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Received unexpected event network-vif-plugged-e54443d3-84a6-4e40-8220-bb2079d7805c for instance with vm_state building and task_state spawning. [ 921.591462] env[63293]: DEBUG nova.compute.manager [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Received event network-changed-e54443d3-84a6-4e40-8220-bb2079d7805c {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.591627] env[63293]: DEBUG nova.compute.manager [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Refreshing instance network info cache due to event network-changed-e54443d3-84a6-4e40-8220-bb2079d7805c. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.591803] env[63293]: DEBUG oslo_concurrency.lockutils [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] Acquiring lock "refresh_cache-2557bded-6ccb-4c0f-8b97-34529b4bb740" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.595844] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327886, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.597401] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "refresh_cache-2557bded-6ccb-4c0f-8b97-34529b4bb740" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.597957] env[63293]: DEBUG nova.compute.manager [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Instance network_info: |[{"id": "e54443d3-84a6-4e40-8220-bb2079d7805c", "address": "fa:16:3e:d3:8e:0c", "network": {"id": "340bfd3b-7dff-454b-87df-788003988adc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-388838680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2af826d819b4cc4ab27ff8b95144039", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape54443d3-84", "ovs_interfaceid": "e54443d3-84a6-4e40-8220-bb2079d7805c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 921.598244] env[63293]: DEBUG oslo_concurrency.lockutils [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] Acquired lock "refresh_cache-2557bded-6ccb-4c0f-8b97-34529b4bb740" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.598514] env[63293]: DEBUG nova.network.neutron [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Refreshing network info cache for port e54443d3-84a6-4e40-8220-bb2079d7805c {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 921.599942] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:8e:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb2eaecd-9701-4504-9fcb-fb1a420ead72', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e54443d3-84a6-4e40-8220-bb2079d7805c', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.608493] env[63293]: DEBUG oslo.service.loopingcall [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.609286] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.610222] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e29591f3-e86f-4f38-bad8-3a47228e00e1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.635040] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.635040] env[63293]: value = "task-1327888" [ 921.635040] env[63293]: _type = "Task" [ 921.635040] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.640187] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327888, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.701861] env[63293]: DEBUG oslo_vmware.api [None req-fb756369-03e8-43bf-b6da-591c456d0130 tempest-ServersAdminNegativeTestJSON-458761815 tempest-ServersAdminNegativeTestJSON-458761815-project-admin] Task: {'id': task-1327887, 'name': SuspendVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.853394] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.855815] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.643s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.857471] env[63293]: INFO nova.compute.claims [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.907395] env[63293]: INFO nova.scheduler.client.report [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted allocations for instance 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d [ 921.948343] env[63293]: DEBUG nova.compute.manager [None req-2289561b-d7af-40f6-b804-732d42bdd33e tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Found 2 images (rotation: 2) {{(pid=63293) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 922.020386] env[63293]: DEBUG nova.compute.manager [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 922.045507] env[63293]: DEBUG nova.virt.hardware [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.046932] env[63293]: DEBUG nova.virt.hardware [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.046932] env[63293]: DEBUG nova.virt.hardware [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.046932] env[63293]: DEBUG nova.virt.hardware [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.046932] env[63293]: DEBUG nova.virt.hardware [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.046932] env[63293]: DEBUG nova.virt.hardware [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.046932] env[63293]: DEBUG nova.virt.hardware [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.046932] env[63293]: DEBUG nova.virt.hardware [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.047485] env[63293]: DEBUG nova.virt.hardware [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.047485] env[63293]: DEBUG nova.virt.hardware [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.047695] env[63293]: DEBUG nova.virt.hardware [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.049033] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d0be9b-62c2-46d5-9ec6-846c34f4d6e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.057799] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a67d87c-22dc-44ed-8a92-2b64753654cf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.085803] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d4dcdb-4c38-56ee-a036-37a877f10120, 'name': SearchDatastore_Task, 'duration_secs': 0.03327} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.089315] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84326892-5a61-4f28-8a3a-b07429eaedaa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.097259] env[63293]: DEBUG oslo_vmware.api [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327886, 'name': PowerOnVM_Task, 'duration_secs': 1.050163} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.097519] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 922.097719] env[63293]: INFO nova.compute.manager [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Took 9.17 seconds to spawn the instance on the hypervisor. [ 922.103199] env[63293]: DEBUG nova.compute.manager [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.103199] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4e3b44-58b2-4e4f-b157-b21bba3e0a67 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.103199] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 922.103199] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c56ea8-b97b-9d4b-5b21-a27b679b7ba5" [ 922.103199] env[63293]: _type = "Task" [ 922.103199] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.115857] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c56ea8-b97b-9d4b-5b21-a27b679b7ba5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.143467] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327888, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.207512] env[63293]: DEBUG oslo_vmware.api [None req-fb756369-03e8-43bf-b6da-591c456d0130 tempest-ServersAdminNegativeTestJSON-458761815 tempest-ServersAdminNegativeTestJSON-458761815-project-admin] Task: {'id': task-1327887, 'name': SuspendVM_Task, 'duration_secs': 1.03286} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.208153] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fb756369-03e8-43bf-b6da-591c456d0130 tempest-ServersAdminNegativeTestJSON-458761815 tempest-ServersAdminNegativeTestJSON-458761815-project-admin] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Suspended the VM {{(pid=63293) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 922.208405] env[63293]: DEBUG nova.compute.manager [None req-fb756369-03e8-43bf-b6da-591c456d0130 tempest-ServersAdminNegativeTestJSON-458761815 tempest-ServersAdminNegativeTestJSON-458761815-project-admin] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.209260] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa318d69-8de3-401d-8074-a75aeb7e4309 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.418618] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ed3044f8-ba92-4b28-81e5-66cead1200eb tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.832s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.487499] env[63293]: DEBUG nova.network.neutron [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Updated VIF entry in instance network info cache for port e54443d3-84a6-4e40-8220-bb2079d7805c. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 922.487725] env[63293]: DEBUG nova.network.neutron [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Updating instance_info_cache with network_info: [{"id": "e54443d3-84a6-4e40-8220-bb2079d7805c", "address": "fa:16:3e:d3:8e:0c", "network": {"id": "340bfd3b-7dff-454b-87df-788003988adc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-388838680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2af826d819b4cc4ab27ff8b95144039", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape54443d3-84", "ovs_interfaceid": "e54443d3-84a6-4e40-8220-bb2079d7805c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.530035] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.530248] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.530502] env[63293]: DEBUG nova.compute.manager [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Going to confirm migration 1 {{(pid=63293) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 922.613107] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c56ea8-b97b-9d4b-5b21-a27b679b7ba5, 'name': SearchDatastore_Task, 'duration_secs': 0.015156} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.613436] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.613743] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] fe55a9b7-28b5-4b04-84f1-6b85b579a144/fe55a9b7-28b5-4b04-84f1-6b85b579a144.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 922.614044] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1fcbc642-17c8-4ad6-bc2b-46846e43ddb4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.631145] env[63293]: INFO nova.compute.manager [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Took 31.59 seconds to build instance. [ 922.634875] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 922.634875] env[63293]: value = "task-1327889" [ 922.634875] env[63293]: _type = "Task" [ 922.634875] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.635265] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.635520] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.653613] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327888, 'name': CreateVM_Task, 'duration_secs': 0.570927} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.654914] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 922.655552] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327889, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.656152] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.656330] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.656732] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.657087] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9517dbe4-b7ff-4fcb-978f-d70cabd0bd9d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.662007] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 922.662007] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529adbea-bff3-508e-9715-38970ed0ad69" [ 922.662007] env[63293]: _type = "Task" [ 922.662007] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.672910] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529adbea-bff3-508e-9715-38970ed0ad69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.990718] env[63293]: DEBUG oslo_concurrency.lockutils [req-aefd99ba-4bba-478d-b56f-20c52727354d req-95ff5be0-4d2f-415b-aa3d-e1e772638221 service nova] Releasing lock "refresh_cache-2557bded-6ccb-4c0f-8b97-34529b4bb740" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.161818] env[63293]: DEBUG oslo_concurrency.lockutils [None req-61029ded-5712-41cc-a8eb-280e182e5d86 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Lock "8625a358-ec48-46b9-8738-c49a2ba58362" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.112s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.161818] env[63293]: DEBUG nova.compute.manager [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.161818] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327889, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.176820] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.179895] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Processing image 8b176c14-05dc-43c1-a475-57283dc96350 {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.179895] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350/8b176c14-05dc-43c1-a475-57283dc96350.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.179895] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350/8b176c14-05dc-43c1-a475-57283dc96350.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.179895] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.179895] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-57dc6721-ef33-46b2-bda3-1322edbb6a37 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.194229] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.194229] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 923.194229] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56bdfadc-b51a-4b7f-ac84-01d09331a534 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.201166] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 923.201166] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520e3cac-d9df-9db0-2fe2-f56aa3887a9a" [ 923.201166] env[63293]: _type = "Task" [ 923.201166] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.213343] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520e3cac-d9df-9db0-2fe2-f56aa3887a9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.249227] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.249422] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.249641] env[63293]: DEBUG nova.network.neutron [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 923.249887] env[63293]: DEBUG nova.objects.instance [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lazy-loading 'info_cache' on Instance uuid 76ebe2ec-332c-4d3a-957e-d0c928dafdc7 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.274480] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0c0808-2104-4bc5-a0e7-a96ce5e2aef3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.282977] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4378701-46b3-49c7-a4ce-3d2e38a0c9c5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.314626] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52d80e8-2313-438d-9ee3-c119d5152a18 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.321904] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c01b6a-2a04-48de-b990-e8a6e46f9db9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.337957] env[63293]: DEBUG nova.compute.manager [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.338480] env[63293]: DEBUG nova.compute.provider_tree [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.340356] env[63293]: DEBUG nova.network.neutron [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Successfully updated port: 4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.343593] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750a6554-19f2-4fba-9314-4fdd60f6882a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.458438] env[63293]: DEBUG nova.compute.manager [req-1ce2af9c-3949-43b7-8e86-61233dc9a673 req-4fc39fca-30b4-48ec-9b1e-af73d0ab273f service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Received event network-vif-plugged-4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.458710] env[63293]: DEBUG oslo_concurrency.lockutils [req-1ce2af9c-3949-43b7-8e86-61233dc9a673 req-4fc39fca-30b4-48ec-9b1e-af73d0ab273f service nova] Acquiring lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.458875] env[63293]: DEBUG oslo_concurrency.lockutils [req-1ce2af9c-3949-43b7-8e86-61233dc9a673 req-4fc39fca-30b4-48ec-9b1e-af73d0ab273f service nova] Lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.459066] env[63293]: DEBUG oslo_concurrency.lockutils [req-1ce2af9c-3949-43b7-8e86-61233dc9a673 req-4fc39fca-30b4-48ec-9b1e-af73d0ab273f service nova] Lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.459251] env[63293]: DEBUG nova.compute.manager [req-1ce2af9c-3949-43b7-8e86-61233dc9a673 req-4fc39fca-30b4-48ec-9b1e-af73d0ab273f service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] No waiting events found dispatching network-vif-plugged-4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.459422] env[63293]: WARNING nova.compute.manager [req-1ce2af9c-3949-43b7-8e86-61233dc9a673 req-4fc39fca-30b4-48ec-9b1e-af73d0ab273f service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Received unexpected event network-vif-plugged-4c817793-6a1c-4d84-ac42-0487913a257b for instance with vm_state building and task_state spawning. [ 923.655237] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327889, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640026} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.655404] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] fe55a9b7-28b5-4b04-84f1-6b85b579a144/fe55a9b7-28b5-4b04-84f1-6b85b579a144.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 923.655641] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 923.657899] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1f6706d3-7b0a-4366-bff8-917dffde1b82 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.665776] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 923.665776] env[63293]: value = "task-1327890" [ 923.665776] env[63293]: _type = "Task" [ 923.665776] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.673685] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327890, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.682421] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.712922] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Preparing fetch location {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 923.713219] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Fetch image to [datastore2] OSTACK_IMG_0f8ee4ff-4755-4acd-bbbc-310031340425/OSTACK_IMG_0f8ee4ff-4755-4acd-bbbc-310031340425.vmdk {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 923.713425] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Downloading stream optimized image 8b176c14-05dc-43c1-a475-57283dc96350 to [datastore2] OSTACK_IMG_0f8ee4ff-4755-4acd-bbbc-310031340425/OSTACK_IMG_0f8ee4ff-4755-4acd-bbbc-310031340425.vmdk on the data store datastore2 as vApp {{(pid=63293) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 923.713703] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Downloading image file data 8b176c14-05dc-43c1-a475-57283dc96350 to the ESX as VM named 'OSTACK_IMG_0f8ee4ff-4755-4acd-bbbc-310031340425' {{(pid=63293) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 923.815795] env[63293]: DEBUG oslo_vmware.rw_handles [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 923.815795] env[63293]: value = "resgroup-9" [ 923.815795] env[63293]: _type = "ResourcePool" [ 923.815795] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 923.816542] env[63293]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-90bd7385-c8d2-4097-b689-c3b2ad2e0797 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.839477] env[63293]: DEBUG oslo_vmware.rw_handles [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lease: (returnval){ [ 923.839477] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d29dd6-bf1d-1b5f-7b9d-c8b99c26a011" [ 923.839477] env[63293]: _type = "HttpNfcLease" [ 923.839477] env[63293]: } obtained for vApp import into resource pool (val){ [ 923.839477] env[63293]: value = "resgroup-9" [ 923.839477] env[63293]: _type = "ResourcePool" [ 923.839477] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 923.839910] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the lease: (returnval){ [ 923.839910] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d29dd6-bf1d-1b5f-7b9d-c8b99c26a011" [ 923.839910] env[63293]: _type = "HttpNfcLease" [ 923.839910] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 923.846069] env[63293]: DEBUG nova.scheduler.client.report [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.850232] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.852225] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.852225] env[63293]: DEBUG nova.network.neutron [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 923.855145] env[63293]: INFO nova.compute.manager [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] instance snapshotting [ 923.855208] env[63293]: DEBUG nova.objects.instance [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'flavor' on Instance uuid 9ddf4f23-1279-4cbd-8212-10f344060445 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.856543] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 923.856543] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d29dd6-bf1d-1b5f-7b9d-c8b99c26a011" [ 923.856543] env[63293]: _type = "HttpNfcLease" [ 923.856543] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 924.175880] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065698} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.176291] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 924.177094] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e197da4-7db1-4977-814f-8ea33260d71e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.201293] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] fe55a9b7-28b5-4b04-84f1-6b85b579a144/fe55a9b7-28b5-4b04-84f1-6b85b579a144.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.201624] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0857fd90-e3c7-4297-aad1-a55a09e7a271 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.223442] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 924.223442] env[63293]: value = "task-1327892" [ 924.223442] env[63293]: _type = "Task" [ 924.223442] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.231968] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327892, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.309705] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "828180c0-b7f8-4666-9f5d-e6741ef6495a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.309995] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "828180c0-b7f8-4666-9f5d-e6741ef6495a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.348128] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 924.348128] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d29dd6-bf1d-1b5f-7b9d-c8b99c26a011" [ 924.348128] env[63293]: _type = "HttpNfcLease" [ 924.348128] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 924.354942] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.355183] env[63293]: DEBUG nova.compute.manager [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 924.357837] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.668s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.358151] env[63293]: DEBUG nova.objects.instance [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lazy-loading 'pci_requests' on Instance uuid ad585ebb-2072-45df-b645-94c9fa93576b {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.364581] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f7a05f-b124-4996-a17e-6b53ff527d38 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.388514] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42efbda4-9eb6-41ed-bb66-0628918ec801 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.410680] env[63293]: DEBUG nova.network.neutron [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 924.500094] env[63293]: DEBUG nova.network.neutron [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance_info_cache with network_info: [{"id": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "address": "fa:16:3e:12:25:eb", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46d5563f-41", "ovs_interfaceid": "46d5563f-41a2-48ab-9d32-3aa394a13da8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.605071] env[63293]: DEBUG nova.network.neutron [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updating instance_info_cache with network_info: [{"id": "4c817793-6a1c-4d84-ac42-0487913a257b", "address": "fa:16:3e:18:84:4b", "network": {"id": "8af42ccf-7ab8-424b-a290-8383c292ef85", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1850566934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eec85c5992d64d62a90e290a1aa2d441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c817793-6a", "ovs_interfaceid": "4c817793-6a1c-4d84-ac42-0487913a257b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.736237] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327892, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.816911] env[63293]: DEBUG nova.compute.manager [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 924.852095] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 924.852095] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d29dd6-bf1d-1b5f-7b9d-c8b99c26a011" [ 924.852095] env[63293]: _type = "HttpNfcLease" [ 924.852095] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 924.865929] env[63293]: DEBUG nova.objects.instance [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lazy-loading 'numa_topology' on Instance uuid ad585ebb-2072-45df-b645-94c9fa93576b {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.868027] env[63293]: DEBUG nova.compute.utils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 924.870020] env[63293]: DEBUG nova.compute.manager [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 924.870232] env[63293]: DEBUG nova.network.neutron [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 924.900120] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Creating Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 924.900446] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-556e35fb-79e6-4771-9421-47190b930121 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.907799] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 924.907799] env[63293]: value = "task-1327893" [ 924.907799] env[63293]: _type = "Task" [ 924.907799] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.917718] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327893, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.919662] env[63293]: DEBUG nova.policy [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2312d48038d7474481a0709267645780', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '866b347100454019b07f63922b995bb6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 925.012102] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-76ebe2ec-332c-4d3a-957e-d0c928dafdc7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.012102] env[63293]: DEBUG nova.objects.instance [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lazy-loading 'migration_context' on Instance uuid 76ebe2ec-332c-4d3a-957e-d0c928dafdc7 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.111183] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.111183] env[63293]: DEBUG nova.compute.manager [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Instance network_info: |[{"id": "4c817793-6a1c-4d84-ac42-0487913a257b", "address": "fa:16:3e:18:84:4b", "network": {"id": "8af42ccf-7ab8-424b-a290-8383c292ef85", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1850566934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eec85c5992d64d62a90e290a1aa2d441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c817793-6a", "ovs_interfaceid": "4c817793-6a1c-4d84-ac42-0487913a257b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 925.111183] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:84:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd48f0ef6-34e5-44d4-8baf-4470ed96ce73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c817793-6a1c-4d84-ac42-0487913a257b', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 925.118314] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Creating folder: Project (eec85c5992d64d62a90e290a1aa2d441). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 925.118981] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6e81334-83b4-480d-8c2d-e144e4039fc7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.131601] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Created folder: Project (eec85c5992d64d62a90e290a1aa2d441) in parent group-v283678. [ 925.131601] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Creating folder: Instances. Parent ref: group-v283801. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 925.132196] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c9b3c1d-86e6-4063-86e2-1e9e7f83d344 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.144131] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Created folder: Instances in parent group-v283801. [ 925.144131] env[63293]: DEBUG oslo.service.loopingcall [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.144131] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 925.144131] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02a91a5c-7ed6-473e-a471-a6c0a103b184 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.167887] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 925.167887] env[63293]: value = "task-1327896" [ 925.167887] env[63293]: _type = "Task" [ 925.167887] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.176263] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327896, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.231023] env[63293]: DEBUG oslo_concurrency.lockutils [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "5b1264b4-2554-4c62-9ff9-4616b03d3609" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.231023] env[63293]: DEBUG oslo_concurrency.lockutils [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "5b1264b4-2554-4c62-9ff9-4616b03d3609" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.231023] env[63293]: DEBUG oslo_concurrency.lockutils [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "5b1264b4-2554-4c62-9ff9-4616b03d3609-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.231023] env[63293]: DEBUG oslo_concurrency.lockutils [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "5b1264b4-2554-4c62-9ff9-4616b03d3609-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.231023] env[63293]: DEBUG oslo_concurrency.lockutils [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "5b1264b4-2554-4c62-9ff9-4616b03d3609-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.234144] env[63293]: INFO nova.compute.manager [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Terminating instance [ 925.242272] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327892, 'name': ReconfigVM_Task, 'duration_secs': 0.674491} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.243227] env[63293]: DEBUG nova.compute.manager [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 925.243620] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 925.244089] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Reconfigured VM instance instance-00000053 to attach disk [datastore1] fe55a9b7-28b5-4b04-84f1-6b85b579a144/fe55a9b7-28b5-4b04-84f1-6b85b579a144.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.245483] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f315b9-ccd5-4ef4-8257-1d48cac2a134 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.248903] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f507e86-89d9-4700-b992-90a8bb34fd84 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.257566] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 925.258461] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8463ba58-6e9e-45fc-8bb8-7c60f28fb58e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.260539] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 925.260539] env[63293]: value = "task-1327897" [ 925.260539] env[63293]: _type = "Task" [ 925.260539] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.271596] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327897, 'name': Rename_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.331318] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.331577] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.331779] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Deleting the datastore file [datastore1] 5b1264b4-2554-4c62-9ff9-4616b03d3609 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.332120] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-428b2cae-c8e3-4fbf-9905-c820950c59ac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.340476] env[63293]: DEBUG oslo_vmware.api [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 925.340476] env[63293]: value = "task-1327899" [ 925.340476] env[63293]: _type = "Task" [ 925.340476] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.348632] env[63293]: DEBUG nova.network.neutron [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Successfully created port: df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.359284] env[63293]: DEBUG oslo_vmware.api [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327899, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.361715] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.364496] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 925.364496] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d29dd6-bf1d-1b5f-7b9d-c8b99c26a011" [ 925.364496] env[63293]: _type = "HttpNfcLease" [ 925.364496] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 925.364923] env[63293]: DEBUG oslo_vmware.rw_handles [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 925.364923] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d29dd6-bf1d-1b5f-7b9d-c8b99c26a011" [ 925.364923] env[63293]: _type = "HttpNfcLease" [ 925.364923] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 925.366192] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9995c1ce-5ea2-479e-a986-2a27cee41338 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.370514] env[63293]: INFO nova.compute.claims [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 925.378020] env[63293]: DEBUG nova.compute.manager [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 925.391285] env[63293]: DEBUG oslo_vmware.rw_handles [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d297b2-983e-7351-215e-102c331fc1f9/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 925.391683] env[63293]: DEBUG oslo_vmware.rw_handles [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d297b2-983e-7351-215e-102c331fc1f9/disk-0.vmdk. {{(pid=63293) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 925.482716] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1f039bcb-28bc-4df3-ac90-6c0b096f32cb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.484868] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327893, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.488173] env[63293]: DEBUG nova.compute.manager [req-d4c370ae-35a9-4ac7-b023-fa7bf136453e req-75138323-fc22-413f-80e3-6ee52e62c942 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Received event network-changed-4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.488389] env[63293]: DEBUG nova.compute.manager [req-d4c370ae-35a9-4ac7-b023-fa7bf136453e req-75138323-fc22-413f-80e3-6ee52e62c942 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Refreshing instance network info cache due to event network-changed-4c817793-6a1c-4d84-ac42-0487913a257b. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.488636] env[63293]: DEBUG oslo_concurrency.lockutils [req-d4c370ae-35a9-4ac7-b023-fa7bf136453e req-75138323-fc22-413f-80e3-6ee52e62c942 service nova] Acquiring lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.488846] env[63293]: DEBUG oslo_concurrency.lockutils [req-d4c370ae-35a9-4ac7-b023-fa7bf136453e req-75138323-fc22-413f-80e3-6ee52e62c942 service nova] Acquired lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.489043] env[63293]: DEBUG nova.network.neutron [req-d4c370ae-35a9-4ac7-b023-fa7bf136453e req-75138323-fc22-413f-80e3-6ee52e62c942 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Refreshing network info cache for port 4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 925.510216] env[63293]: DEBUG nova.objects.base [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Object Instance<76ebe2ec-332c-4d3a-957e-d0c928dafdc7> lazy-loaded attributes: info_cache,migration_context {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 925.512552] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f90050-666a-4707-b6f7-ab70cef4e08d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.533806] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ac79801-2cc5-47e6-bb13-47fd95a2da57 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.539556] env[63293]: DEBUG oslo_vmware.api [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 925.539556] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5272ebfe-3840-e61a-3b2c-76a9de1357ca" [ 925.539556] env[63293]: _type = "Task" [ 925.539556] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.547195] env[63293]: DEBUG oslo_vmware.api [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5272ebfe-3840-e61a-3b2c-76a9de1357ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.680287] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327896, 'name': CreateVM_Task, 'duration_secs': 0.348717} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.680481] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 925.681229] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.681405] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.681751] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 925.682038] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a8f7018-92d3-4770-8dc3-34dc7952b0cc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.686857] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 925.686857] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52355c8a-a385-b4a9-c5a2-2c3688ef1990" [ 925.686857] env[63293]: _type = "Task" [ 925.686857] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.694462] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52355c8a-a385-b4a9-c5a2-2c3688ef1990, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.770459] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327897, 'name': Rename_Task, 'duration_secs': 0.156243} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.770737] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.770945] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd38f05b-2771-4254-b928-e87521503c22 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.777887] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquiring lock "8625a358-ec48-46b9-8738-c49a2ba58362" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.778159] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Lock "8625a358-ec48-46b9-8738-c49a2ba58362" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.778373] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquiring lock "8625a358-ec48-46b9-8738-c49a2ba58362-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.778565] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Lock "8625a358-ec48-46b9-8738-c49a2ba58362-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.778736] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Lock "8625a358-ec48-46b9-8738-c49a2ba58362-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.780464] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 925.780464] env[63293]: value = "task-1327900" [ 925.780464] env[63293]: _type = "Task" [ 925.780464] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.780964] env[63293]: INFO nova.compute.manager [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Terminating instance [ 925.785252] env[63293]: DEBUG nova.compute.manager [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 925.785448] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 925.786198] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4f91e1-2306-4af9-941b-b853850ca801 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.794062] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327900, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.796131] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.796382] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4baa333-968d-4c8a-9d43-c83d8f266f4a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.802644] env[63293]: DEBUG oslo_vmware.api [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for the task: (returnval){ [ 925.802644] env[63293]: value = "task-1327901" [ 925.802644] env[63293]: _type = "Task" [ 925.802644] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.810872] env[63293]: DEBUG oslo_vmware.api [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327901, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.849758] env[63293]: DEBUG oslo_vmware.api [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327899, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190991} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.850248] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.850248] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.850417] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.850589] env[63293]: INFO nova.compute.manager [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Took 0.61 seconds to destroy the instance on the hypervisor. [ 925.850841] env[63293]: DEBUG oslo.service.loopingcall [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.851049] env[63293]: DEBUG nova.compute.manager [-] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 925.851146] env[63293]: DEBUG nova.network.neutron [-] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 925.983261] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327893, 'name': CreateSnapshot_Task, 'duration_secs': 0.586549} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.983561] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Created Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 925.984334] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e108a02a-c5ba-452c-ba0b-bc06806da0d7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.051810] env[63293]: DEBUG oslo_vmware.api [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5272ebfe-3840-e61a-3b2c-76a9de1357ca, 'name': SearchDatastore_Task, 'duration_secs': 0.017629} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.052132] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.173491] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9641ffaf-d181-4d11-a390-6d74930442d2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.197308] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f05aa3ba-2052-403c-8923-343ba87f0ff7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.242640] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52355c8a-a385-b4a9-c5a2-2c3688ef1990, 'name': SearchDatastore_Task, 'duration_secs': 0.028534} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.251026] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.254022] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.254022] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.254022] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.254022] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.254022] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7f83e8-a202-4706-998a-9c1f934f34ce {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.258453] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30e07e71-6f9f-42d2-ac73-6cd803664459 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.267856] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b704d707-71c6-46bf-87a7-02f68a6cfeb3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.274325] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.274535] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 926.277036] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-061d4bd5-940d-454f-89f8-41a30b09b37d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.293665] env[63293]: DEBUG nova.compute.provider_tree [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.305651] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 926.305651] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5291d3e2-489a-709a-6367-f6fd6928a40c" [ 926.305651] env[63293]: _type = "Task" [ 926.305651] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.316193] env[63293]: DEBUG oslo_vmware.api [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327900, 'name': PowerOnVM_Task, 'duration_secs': 0.530393} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.321906] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 926.322130] env[63293]: INFO nova.compute.manager [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Took 9.13 seconds to spawn the instance on the hypervisor. [ 926.322318] env[63293]: DEBUG nova.compute.manager [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.323709] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3c4eca-1711-48bd-aa83-d4b660f37f62 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.334434] env[63293]: DEBUG oslo_vmware.api [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327901, 'name': PowerOffVM_Task, 'duration_secs': 0.272127} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.334677] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5291d3e2-489a-709a-6367-f6fd6928a40c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.335315] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.335493] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 926.335752] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e213bfe-42dc-45f5-bad6-a9232cef9e47 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.387991] env[63293]: DEBUG nova.compute.manager [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 926.430565] env[63293]: DEBUG nova.virt.hardware [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.430875] env[63293]: DEBUG nova.virt.hardware [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.431094] env[63293]: DEBUG nova.virt.hardware [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.431297] env[63293]: DEBUG nova.virt.hardware [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.431449] env[63293]: DEBUG nova.virt.hardware [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.431600] env[63293]: DEBUG nova.virt.hardware [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.431811] env[63293]: DEBUG nova.virt.hardware [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.431978] env[63293]: DEBUG nova.virt.hardware [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.432165] env[63293]: DEBUG nova.virt.hardware [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.432352] env[63293]: DEBUG nova.virt.hardware [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.432559] env[63293]: DEBUG nova.virt.hardware [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.433444] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401171bb-2873-43a3-a116-f49321f21369 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.438399] env[63293]: DEBUG nova.network.neutron [req-d4c370ae-35a9-4ac7-b023-fa7bf136453e req-75138323-fc22-413f-80e3-6ee52e62c942 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updated VIF entry in instance network info cache for port 4c817793-6a1c-4d84-ac42-0487913a257b. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 926.438739] env[63293]: DEBUG nova.network.neutron [req-d4c370ae-35a9-4ac7-b023-fa7bf136453e req-75138323-fc22-413f-80e3-6ee52e62c942 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updating instance_info_cache with network_info: [{"id": "4c817793-6a1c-4d84-ac42-0487913a257b", "address": "fa:16:3e:18:84:4b", "network": {"id": "8af42ccf-7ab8-424b-a290-8383c292ef85", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1850566934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eec85c5992d64d62a90e290a1aa2d441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c817793-6a", "ovs_interfaceid": "4c817793-6a1c-4d84-ac42-0487913a257b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.445347] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc09229-76d4-4d2b-9d39-a12080d5eb2f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.516503] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Creating linked-clone VM from snapshot {{(pid=63293) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 926.516936] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a3b0046b-757a-4bba-888e-7cac5af6f452 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.527831] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 926.527831] env[63293]: value = "task-1327903" [ 926.527831] env[63293]: _type = "Task" [ 926.527831] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.539987] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327903, 'name': CloneVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.649377] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 926.649671] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 926.649925] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Deleting the datastore file [datastore1] 8625a358-ec48-46b9-8738-c49a2ba58362 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 926.650297] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad5683dd-56de-46ce-a26c-e2abd2e06bd6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.658045] env[63293]: DEBUG oslo_vmware.api [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for the task: (returnval){ [ 926.658045] env[63293]: value = "task-1327904" [ 926.658045] env[63293]: _type = "Task" [ 926.658045] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.669166] env[63293]: DEBUG oslo_vmware.api [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327904, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.781356] env[63293]: DEBUG nova.network.neutron [-] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.806252] env[63293]: DEBUG nova.scheduler.client.report [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.819792] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5291d3e2-489a-709a-6367-f6fd6928a40c, 'name': SearchDatastore_Task, 'duration_secs': 0.03079} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.820742] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-653385be-1ec5-430b-9adf-0ec00aeab74b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.826768] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 926.826768] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522b1ae8-df0d-a38f-9557-d35138ad2694" [ 926.826768] env[63293]: _type = "Task" [ 926.826768] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.835194] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522b1ae8-df0d-a38f-9557-d35138ad2694, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.854224] env[63293]: INFO nova.compute.manager [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Took 24.35 seconds to build instance. [ 926.941686] env[63293]: DEBUG oslo_concurrency.lockutils [req-d4c370ae-35a9-4ac7-b023-fa7bf136453e req-75138323-fc22-413f-80e3-6ee52e62c942 service nova] Releasing lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.951864] env[63293]: DEBUG nova.compute.manager [req-5f1e3001-6631-4ee9-a262-41086c0a93f1 req-3ae5e420-19de-4584-8ec3-1737f100e0c9 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Received event network-vif-plugged-df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.952107] env[63293]: DEBUG oslo_concurrency.lockutils [req-5f1e3001-6631-4ee9-a262-41086c0a93f1 req-3ae5e420-19de-4584-8ec3-1737f100e0c9 service nova] Acquiring lock "35cbf277-b85a-4101-beae-0125fb5a61f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.952335] env[63293]: DEBUG oslo_concurrency.lockutils [req-5f1e3001-6631-4ee9-a262-41086c0a93f1 req-3ae5e420-19de-4584-8ec3-1737f100e0c9 service nova] Lock "35cbf277-b85a-4101-beae-0125fb5a61f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.952542] env[63293]: DEBUG oslo_concurrency.lockutils [req-5f1e3001-6631-4ee9-a262-41086c0a93f1 req-3ae5e420-19de-4584-8ec3-1737f100e0c9 service nova] Lock "35cbf277-b85a-4101-beae-0125fb5a61f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.952723] env[63293]: DEBUG nova.compute.manager [req-5f1e3001-6631-4ee9-a262-41086c0a93f1 req-3ae5e420-19de-4584-8ec3-1737f100e0c9 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] No waiting events found dispatching network-vif-plugged-df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 926.952886] env[63293]: WARNING nova.compute.manager [req-5f1e3001-6631-4ee9-a262-41086c0a93f1 req-3ae5e420-19de-4584-8ec3-1737f100e0c9 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Received unexpected event network-vif-plugged-df1c83c6-23d7-439b-a496-234c48738d81 for instance with vm_state building and task_state spawning. [ 927.024753] env[63293]: DEBUG oslo_vmware.rw_handles [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Completed reading data from the image iterator. {{(pid=63293) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 927.025070] env[63293]: DEBUG oslo_vmware.rw_handles [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d297b2-983e-7351-215e-102c331fc1f9/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 927.026053] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79809ace-98f4-4f4d-b07d-36681abde53d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.036494] env[63293]: DEBUG oslo_vmware.rw_handles [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d297b2-983e-7351-215e-102c331fc1f9/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 927.036661] env[63293]: DEBUG oslo_vmware.rw_handles [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d297b2-983e-7351-215e-102c331fc1f9/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 927.042522] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-7684dd40-bbbb-4243-b130-43e1c9f9ac26 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.043516] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327903, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.050987] env[63293]: DEBUG nova.network.neutron [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Successfully updated port: df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 927.169026] env[63293]: DEBUG oslo_vmware.api [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Task: {'id': task-1327904, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249846} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.169026] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.169026] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.169267] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.169599] env[63293]: INFO nova.compute.manager [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Took 1.38 seconds to destroy the instance on the hypervisor. [ 927.169768] env[63293]: DEBUG oslo.service.loopingcall [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.170013] env[63293]: DEBUG nova.compute.manager [-] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.170145] env[63293]: DEBUG nova.network.neutron [-] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 927.284207] env[63293]: INFO nova.compute.manager [-] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Took 1.43 seconds to deallocate network for instance. [ 927.315315] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.957s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.318120] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.636s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.320342] env[63293]: INFO nova.compute.claims [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 927.339506] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522b1ae8-df0d-a38f-9557-d35138ad2694, 'name': SearchDatastore_Task, 'duration_secs': 0.028873} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.339776] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.340082] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] f429869f-9e9f-443e-a980-67a80c3bd799/f429869f-9e9f-443e-a980-67a80c3bd799.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 927.340347] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08567209-f94c-4ec1-bb78-4a34e8282bbc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.346728] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 927.346728] env[63293]: value = "task-1327905" [ 927.346728] env[63293]: _type = "Task" [ 927.346728] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.354674] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327905, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.356165] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30299e0d-bcbf-400f-8ebd-0c9f6efa0bca tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.864s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.369538] env[63293]: INFO nova.network.neutron [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updating port 0576b708-1b9b-4159-8079-9b97d8b4e70d with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 927.400787] env[63293]: DEBUG oslo_vmware.rw_handles [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d297b2-983e-7351-215e-102c331fc1f9/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 927.401087] env[63293]: INFO nova.virt.vmwareapi.images [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Downloaded image file data 8b176c14-05dc-43c1-a475-57283dc96350 [ 927.401960] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-060c1766-348c-42b2-960c-f94aeff2f710 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.428026] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05cb186d-2ec6-4dbe-9c78-e1505203ce9e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.450857] env[63293]: INFO nova.virt.vmwareapi.images [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] The imported VM was unregistered [ 927.454123] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Caching image {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 927.454364] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Creating directory with path [datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350 {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.454648] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ea7557a-7b54-4b41-ac37-a4ea0d1b29ac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.466385] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Created directory with path [datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350 {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.467265] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_0f8ee4ff-4755-4acd-bbbc-310031340425/OSTACK_IMG_0f8ee4ff-4755-4acd-bbbc-310031340425.vmdk to [datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350/8b176c14-05dc-43c1-a475-57283dc96350.vmdk. {{(pid=63293) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 927.467265] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-0d60d177-6c37-4ce7-b0ec-7d51bebdf08c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.475338] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 927.475338] env[63293]: value = "task-1327907" [ 927.475338] env[63293]: _type = "Task" [ 927.475338] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.484509] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327907, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.542875] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327903, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.555538] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.555538] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.555538] env[63293]: DEBUG nova.network.neutron [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.587537] env[63293]: DEBUG nova.compute.manager [req-82d628a9-7977-4d8b-8c51-4f9086eef06c req-d1028e82-8270-4391-90dc-6a7adc40f2b7 service nova] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Received event network-vif-deleted-f6501e07-0e63-4597-86b9-bbe243fabaa3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.792977] env[63293]: DEBUG oslo_concurrency.lockutils [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.858560] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327905, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.965555] env[63293]: DEBUG nova.network.neutron [-] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.993688] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327907, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.046660] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327903, 'name': CloneVM_Task} progress is 95%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.112059] env[63293]: DEBUG nova.network.neutron [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 928.177480] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.177917] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.178166] env[63293]: DEBUG nova.compute.manager [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.179185] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-060bc496-e0ac-47f1-93a5-c914bcd06684 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.189934] env[63293]: DEBUG nova.compute.manager [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63293) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 928.190651] env[63293]: DEBUG nova.objects.instance [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lazy-loading 'flavor' on Instance uuid fe55a9b7-28b5-4b04-84f1-6b85b579a144 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.287230] env[63293]: DEBUG nova.network.neutron [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updating instance_info_cache with network_info: [{"id": "df1c83c6-23d7-439b-a496-234c48738d81", "address": "fa:16:3e:0f:9f:14", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1c83c6-23", "ovs_interfaceid": "df1c83c6-23d7-439b-a496-234c48738d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.365727] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327905, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543696} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.368728] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] f429869f-9e9f-443e-a980-67a80c3bd799/f429869f-9e9f-443e-a980-67a80c3bd799.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 928.369021] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.370127] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5fc3ef67-603f-4d90-826f-26c4c57b9fd9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.383958] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 928.383958] env[63293]: value = "task-1327908" [ 928.383958] env[63293]: _type = "Task" [ 928.383958] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.397527] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327908, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.469495] env[63293]: INFO nova.compute.manager [-] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Took 1.30 seconds to deallocate network for instance. [ 928.490677] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327907, 'name': MoveVirtualDisk_Task} progress is 29%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.549486] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327903, 'name': CloneVM_Task, 'duration_secs': 1.875887} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.549890] env[63293]: INFO nova.virt.vmwareapi.vmops [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Created linked-clone VM from snapshot [ 928.550755] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb031caa-a7c8-4ef5-a61a-0a8a93648581 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.563491] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Uploading image 66bcd27d-468f-428b-b4cb-6fed31de06c3 {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 928.580971] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1db3fa8-1a46-406e-98c3-c8d3eb0e16e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.592700] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ef5e18-87b9-4087-9758-92ce07f2a987 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.598122] env[63293]: DEBUG oslo_vmware.rw_handles [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 928.598122] env[63293]: value = "vm-283805" [ 928.598122] env[63293]: _type = "VirtualMachine" [ 928.598122] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 928.598425] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b6eecc57-9733-470e-b85e-b1757b87f78f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.606771] env[63293]: DEBUG oslo_vmware.rw_handles [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lease: (returnval){ [ 928.606771] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c97a36-9a94-589c-433d-2abd8efc37f0" [ 928.606771] env[63293]: _type = "HttpNfcLease" [ 928.606771] env[63293]: } obtained for exporting VM: (result){ [ 928.606771] env[63293]: value = "vm-283805" [ 928.606771] env[63293]: _type = "VirtualMachine" [ 928.606771] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 928.607186] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the lease: (returnval){ [ 928.607186] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c97a36-9a94-589c-433d-2abd8efc37f0" [ 928.607186] env[63293]: _type = "HttpNfcLease" [ 928.607186] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 928.638866] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7728f3-4a7b-4d09-9546-e0387a9a942b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.643107] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 928.643107] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c97a36-9a94-589c-433d-2abd8efc37f0" [ 928.643107] env[63293]: _type = "HttpNfcLease" [ 928.643107] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 928.643773] env[63293]: DEBUG oslo_vmware.rw_handles [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 928.643773] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c97a36-9a94-589c-433d-2abd8efc37f0" [ 928.643773] env[63293]: _type = "HttpNfcLease" [ 928.643773] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 928.644866] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235365bf-5f3a-48d3-b255-7b504becac4d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.654354] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47088416-b72e-4a0c-8af6-47080bfb7cf1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.658927] env[63293]: DEBUG oslo_vmware.rw_handles [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52161e4e-e143-4795-e91e-2a2ab4ab2a5c/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 928.659202] env[63293]: DEBUG oslo_vmware.rw_handles [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52161e4e-e143-4795-e91e-2a2ab4ab2a5c/disk-0.vmdk for reading. {{(pid=63293) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 928.732332] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 928.734336] env[63293]: DEBUG nova.compute.provider_tree [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.735785] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e05e981b-b31a-4733-86a3-e5ea04c2fd1c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.744528] env[63293]: DEBUG oslo_vmware.api [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 928.744528] env[63293]: value = "task-1327910" [ 928.744528] env[63293]: _type = "Task" [ 928.744528] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.753866] env[63293]: DEBUG oslo_vmware.api [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327910, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.789981] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.790349] env[63293]: DEBUG nova.compute.manager [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Instance network_info: |[{"id": "df1c83c6-23d7-439b-a496-234c48738d81", "address": "fa:16:3e:0f:9f:14", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1c83c6-23", "ovs_interfaceid": "df1c83c6-23d7-439b-a496-234c48738d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 928.790915] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:9f:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '459b8c74-0aa6-42b6-996a-42b1c5d7e5c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df1c83c6-23d7-439b-a496-234c48738d81', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.798483] env[63293]: DEBUG oslo.service.loopingcall [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.798746] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.798992] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7c632cd8-bbb0-4ea8-9352-92c764ef0733 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.820027] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.820027] env[63293]: value = "task-1327911" [ 928.820027] env[63293]: _type = "Task" [ 928.820027] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.829535] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327911, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.839043] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-32ba77a8-8cbc-4234-aa25-e17d8dcf601f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.894610] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327908, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.20728} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.894987] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 928.895942] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc50d29-666e-4710-8247-a712fb6a0c19 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.922680] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] f429869f-9e9f-443e-a980-67a80c3bd799/f429869f-9e9f-443e-a980-67a80c3bd799.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.923069] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac461bc0-eb34-4467-b4d5-7f39b5887da1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.944194] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 928.944194] env[63293]: value = "task-1327912" [ 928.944194] env[63293]: _type = "Task" [ 928.944194] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.953891] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327912, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.978831] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.988383] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327907, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.991352] env[63293]: DEBUG nova.compute.manager [req-2f8b89f4-06fd-4c7c-83fb-57bb380b02d0 req-6ebe5fea-d33a-4e95-9307-850fdc7a605e service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Received event network-changed-df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.991528] env[63293]: DEBUG nova.compute.manager [req-2f8b89f4-06fd-4c7c-83fb-57bb380b02d0 req-6ebe5fea-d33a-4e95-9307-850fdc7a605e service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing instance network info cache due to event network-changed-df1c83c6-23d7-439b-a496-234c48738d81. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 928.991749] env[63293]: DEBUG oslo_concurrency.lockutils [req-2f8b89f4-06fd-4c7c-83fb-57bb380b02d0 req-6ebe5fea-d33a-4e95-9307-850fdc7a605e service nova] Acquiring lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.991887] env[63293]: DEBUG oslo_concurrency.lockutils [req-2f8b89f4-06fd-4c7c-83fb-57bb380b02d0 req-6ebe5fea-d33a-4e95-9307-850fdc7a605e service nova] Acquired lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.992051] env[63293]: DEBUG nova.network.neutron [req-2f8b89f4-06fd-4c7c-83fb-57bb380b02d0 req-6ebe5fea-d33a-4e95-9307-850fdc7a605e service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing network info cache for port df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.127047] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.127354] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.127637] env[63293]: DEBUG nova.network.neutron [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 929.239584] env[63293]: DEBUG nova.scheduler.client.report [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.255589] env[63293]: DEBUG oslo_vmware.api [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327910, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.331554] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327911, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.454968] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327912, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.487804] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327907, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.628542] env[63293]: DEBUG nova.compute.manager [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Received event network-vif-plugged-0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.628868] env[63293]: DEBUG oslo_concurrency.lockutils [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] Acquiring lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.629207] env[63293]: DEBUG oslo_concurrency.lockutils [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] Lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.629474] env[63293]: DEBUG oslo_concurrency.lockutils [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] Lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.629865] env[63293]: DEBUG nova.compute.manager [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] No waiting events found dispatching network-vif-plugged-0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.630254] env[63293]: WARNING nova.compute.manager [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Received unexpected event network-vif-plugged-0576b708-1b9b-4159-8079-9b97d8b4e70d for instance with vm_state shelved_offloaded and task_state spawning. [ 929.630519] env[63293]: DEBUG nova.compute.manager [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Received event network-changed-0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.630847] env[63293]: DEBUG nova.compute.manager [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Refreshing instance network info cache due to event network-changed-0576b708-1b9b-4159-8079-9b97d8b4e70d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.631138] env[63293]: DEBUG oslo_concurrency.lockutils [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] Acquiring lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.747025] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.747025] env[63293]: DEBUG nova.compute.manager [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 929.751983] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.391s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.753608] env[63293]: INFO nova.compute.claims [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.771336] env[63293]: DEBUG oslo_vmware.api [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327910, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.801037] env[63293]: DEBUG nova.network.neutron [req-2f8b89f4-06fd-4c7c-83fb-57bb380b02d0 req-6ebe5fea-d33a-4e95-9307-850fdc7a605e service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updated VIF entry in instance network info cache for port df1c83c6-23d7-439b-a496-234c48738d81. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 929.801531] env[63293]: DEBUG nova.network.neutron [req-2f8b89f4-06fd-4c7c-83fb-57bb380b02d0 req-6ebe5fea-d33a-4e95-9307-850fdc7a605e service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updating instance_info_cache with network_info: [{"id": "df1c83c6-23d7-439b-a496-234c48738d81", "address": "fa:16:3e:0f:9f:14", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1c83c6-23", "ovs_interfaceid": "df1c83c6-23d7-439b-a496-234c48738d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.832742] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327911, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.957936] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327912, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.990773] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327907, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.030242] env[63293]: DEBUG nova.network.neutron [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updating instance_info_cache with network_info: [{"id": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "address": "fa:16:3e:c0:e8:15", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0576b708-1b", "ovs_interfaceid": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.261506] env[63293]: DEBUG nova.compute.utils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.263124] env[63293]: DEBUG nova.compute.manager [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 930.263264] env[63293]: DEBUG nova.network.neutron [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 930.272231] env[63293]: DEBUG oslo_vmware.api [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327910, 'name': PowerOffVM_Task, 'duration_secs': 1.083251} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.273041] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.273254] env[63293]: DEBUG nova.compute.manager [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.274220] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce186737-1b1d-4383-961a-2fd0f70180af {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.305166] env[63293]: DEBUG oslo_concurrency.lockutils [req-2f8b89f4-06fd-4c7c-83fb-57bb380b02d0 req-6ebe5fea-d33a-4e95-9307-850fdc7a605e service nova] Releasing lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.305395] env[63293]: DEBUG nova.compute.manager [req-2f8b89f4-06fd-4c7c-83fb-57bb380b02d0 req-6ebe5fea-d33a-4e95-9307-850fdc7a605e service nova] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Received event network-vif-deleted-9551ebb5-552d-47d0-a436-994fe2b1d00b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.315215] env[63293]: DEBUG nova.policy [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7a950baa92f4d9ea77e5d6461582e0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec53642df3804e7190615487a426d4a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 930.331876] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327911, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.457369] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327912, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.491235] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327907, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.874242} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.491538] env[63293]: INFO nova.virt.vmwareapi.ds_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_0f8ee4ff-4755-4acd-bbbc-310031340425/OSTACK_IMG_0f8ee4ff-4755-4acd-bbbc-310031340425.vmdk to [datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350/8b176c14-05dc-43c1-a475-57283dc96350.vmdk. [ 930.491761] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Cleaning up location [datastore2] OSTACK_IMG_0f8ee4ff-4755-4acd-bbbc-310031340425 {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 930.492030] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_0f8ee4ff-4755-4acd-bbbc-310031340425 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.492429] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc0f7c5e-8b7a-4837-b5bc-32a175694147 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.498867] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 930.498867] env[63293]: value = "task-1327913" [ 930.498867] env[63293]: _type = "Task" [ 930.498867] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.506796] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327913, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.533616] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.536150] env[63293]: DEBUG oslo_concurrency.lockutils [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] Acquired lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.536345] env[63293]: DEBUG nova.network.neutron [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Refreshing network info cache for port 0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 930.562782] env[63293]: DEBUG nova.virt.hardware [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='39cfee56b784b5ca3ffe365f743bc58d',container_format='bare',created_at=2024-10-21T17:39:39Z,direct_url=,disk_format='vmdk',id=dcac372d-af95-4395-b80b-840bcd2b2ed4,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-348291096-shelved',owner='affba61d6a2846b38666544bc2c25db5',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-21T17:39:56Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.563187] env[63293]: DEBUG nova.virt.hardware [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.563386] env[63293]: DEBUG nova.virt.hardware [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.563629] env[63293]: DEBUG nova.virt.hardware [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.563836] env[63293]: DEBUG nova.virt.hardware [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.564077] env[63293]: DEBUG nova.virt.hardware [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.564357] env[63293]: DEBUG nova.virt.hardware [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.564720] env[63293]: DEBUG nova.virt.hardware [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.565378] env[63293]: DEBUG nova.virt.hardware [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.565378] env[63293]: DEBUG nova.virt.hardware [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.566754] env[63293]: DEBUG nova.virt.hardware [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.567432] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082e1882-1b59-483d-94d2-a3b23a38280f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.575579] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aad0cfc-8274-4d84-8781-be01f37950e6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.592141] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:e8:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24144f5a-050a-4f1e-8d8c-774dc16dc791', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0576b708-1b9b-4159-8079-9b97d8b4e70d', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.605956] env[63293]: DEBUG oslo.service.loopingcall [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.606438] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 930.606746] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6b3ef4a-dcb2-4a23-a467-fbfae9e04cc9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.625430] env[63293]: DEBUG nova.network.neutron [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Successfully created port: 9e8240fa-3546-43eb-9cdc-4b42859690b3 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 930.629323] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.629323] env[63293]: value = "task-1327914" [ 930.629323] env[63293]: _type = "Task" [ 930.629323] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.638894] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327914, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.766208] env[63293]: DEBUG nova.compute.manager [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 930.793556] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f510a5c3-23a6-4d10-81ec-8d2ee0e8551c tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.616s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.836687] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327911, 'name': CreateVM_Task, 'duration_secs': 1.974689} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.836907] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 930.837744] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.838273] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.838742] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.841881] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45c066f6-5706-4e7b-a110-c671f8eb3294 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.847991] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 930.847991] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525ae845-4315-e7ab-eb34-04ca8f9af3f1" [ 930.847991] env[63293]: _type = "Task" [ 930.847991] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.860689] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525ae845-4315-e7ab-eb34-04ca8f9af3f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.962537] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327912, 'name': ReconfigVM_Task, 'duration_secs': 1.620571} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.963249] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Reconfigured VM instance instance-00000055 to attach disk [datastore2] f429869f-9e9f-443e-a980-67a80c3bd799/f429869f-9e9f-443e-a980-67a80c3bd799.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.964192] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-332889e5-2627-4ec1-9316-33fed042ad9b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.974619] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 930.974619] env[63293]: value = "task-1327915" [ 930.974619] env[63293]: _type = "Task" [ 930.974619] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.985547] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327915, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.011650] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327913, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090673} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.012279] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.012279] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350/8b176c14-05dc-43c1-a475-57283dc96350.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.013719] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350/8b176c14-05dc-43c1-a475-57283dc96350.vmdk to [datastore2] 2557bded-6ccb-4c0f-8b97-34529b4bb740/2557bded-6ccb-4c0f-8b97-34529b4bb740.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 931.015913] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-975d7f51-26a1-4641-b1a7-5eb54dbf7eeb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.024721] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 931.024721] env[63293]: value = "task-1327916" [ 931.024721] env[63293]: _type = "Task" [ 931.024721] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.035685] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327916, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.064978] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3522e8f7-a160-42dc-8b2d-e146452c4a13 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.072869] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d82c38a-ca8a-407e-8c9c-cb963ee29285 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.106615] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf63381e-2bd0-4f7a-94c3-a15ccf4e8de5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.115368] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9391d14f-dffd-4d08-a121-a7907be2be26 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.132452] env[63293]: DEBUG nova.compute.provider_tree [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.142797] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327914, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.346216] env[63293]: DEBUG nova.network.neutron [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updated VIF entry in instance network info cache for port 0576b708-1b9b-4159-8079-9b97d8b4e70d. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 931.346720] env[63293]: DEBUG nova.network.neutron [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updating instance_info_cache with network_info: [{"id": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "address": "fa:16:3e:c0:e8:15", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0576b708-1b", "ovs_interfaceid": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.360834] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525ae845-4315-e7ab-eb34-04ca8f9af3f1, 'name': SearchDatastore_Task, 'duration_secs': 0.019173} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.361671] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.361671] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.361978] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.362144] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.363250] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.363250] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-856a18c9-549d-4b44-beaa-62ed8f7790f3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.379625] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.380045] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 931.381497] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d5da89b-7779-4b92-93cf-d5b5f5c28b00 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.388139] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 931.388139] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5224b9c3-cc7b-5ce0-2dda-077031f7ce58" [ 931.388139] env[63293]: _type = "Task" [ 931.388139] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.395806] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5224b9c3-cc7b-5ce0-2dda-077031f7ce58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.485987] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327915, 'name': Rename_Task, 'duration_secs': 0.228825} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.486356] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.486628] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-889ed311-3a1f-4905-b195-aca124b72c48 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.494835] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 931.494835] env[63293]: value = "task-1327917" [ 931.494835] env[63293]: _type = "Task" [ 931.494835] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.503755] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327917, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.534763] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327916, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.639200] env[63293]: DEBUG nova.scheduler.client.report [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.653020] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327914, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.775729] env[63293]: DEBUG nova.compute.manager [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 931.803204] env[63293]: DEBUG nova.virt.hardware [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 931.803531] env[63293]: DEBUG nova.virt.hardware [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 931.803698] env[63293]: DEBUG nova.virt.hardware [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.803899] env[63293]: DEBUG nova.virt.hardware [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 931.804059] env[63293]: DEBUG nova.virt.hardware [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.804215] env[63293]: DEBUG nova.virt.hardware [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 931.804433] env[63293]: DEBUG nova.virt.hardware [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 931.804598] env[63293]: DEBUG nova.virt.hardware [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 931.804778] env[63293]: DEBUG nova.virt.hardware [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 931.804930] env[63293]: DEBUG nova.virt.hardware [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 931.805440] env[63293]: DEBUG nova.virt.hardware [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.806382] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c13c2d2d-e05b-414e-8716-126d18017f0a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.814719] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b02f32c-b038-416a-b0de-5262c3c6dfd2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.819154] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.819398] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.819618] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.820019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.820225] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.822578] env[63293]: INFO nova.compute.manager [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Terminating instance [ 931.834701] env[63293]: DEBUG nova.compute.manager [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 931.834934] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 931.835755] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cda6600-c151-4c45-8033-0899c4600eb6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.843505] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 931.843781] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab93af7a-27f2-42e9-b60a-d919c8622eae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.849120] env[63293]: DEBUG oslo_concurrency.lockutils [req-f31e0af0-71d4-41a0-a396-78340d0de286 req-2b5ec244-9dc4-4478-a6c5-33d8db7d4785 service nova] Releasing lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.901913] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5224b9c3-cc7b-5ce0-2dda-077031f7ce58, 'name': SearchDatastore_Task, 'duration_secs': 0.076666} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.902759] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4da1c3ca-b8d8-475c-a20e-003f400858e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.911254] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 931.911254] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5259f93b-cdb9-6c96-b9c5-fd45d121b2c5" [ 931.911254] env[63293]: _type = "Task" [ 931.911254] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.922762] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 931.922995] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 931.923199] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleting the datastore file [datastore1] fe55a9b7-28b5-4b04-84f1-6b85b579a144 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.926720] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc1a15d6-c7a4-40bf-baa9-7f30f99257b5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.928714] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5259f93b-cdb9-6c96-b9c5-fd45d121b2c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.935210] env[63293]: DEBUG oslo_vmware.api [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 931.935210] env[63293]: value = "task-1327919" [ 931.935210] env[63293]: _type = "Task" [ 931.935210] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.945596] env[63293]: DEBUG oslo_vmware.api [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327919, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.008136] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327917, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.038556] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327916, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.147114] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.147760] env[63293]: DEBUG nova.compute.manager [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.150499] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327914, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.150822] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 6.099s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.286186] env[63293]: DEBUG nova.compute.manager [req-25d535af-a68c-43b0-97d3-e559672eee54 req-6f887816-7694-468c-9d3c-21b30de5f0f3 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Received event network-vif-plugged-9e8240fa-3546-43eb-9cdc-4b42859690b3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.288418] env[63293]: DEBUG oslo_concurrency.lockutils [req-25d535af-a68c-43b0-97d3-e559672eee54 req-6f887816-7694-468c-9d3c-21b30de5f0f3 service nova] Acquiring lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.288757] env[63293]: DEBUG oslo_concurrency.lockutils [req-25d535af-a68c-43b0-97d3-e559672eee54 req-6f887816-7694-468c-9d3c-21b30de5f0f3 service nova] Lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.289042] env[63293]: DEBUG oslo_concurrency.lockutils [req-25d535af-a68c-43b0-97d3-e559672eee54 req-6f887816-7694-468c-9d3c-21b30de5f0f3 service nova] Lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.289314] env[63293]: DEBUG nova.compute.manager [req-25d535af-a68c-43b0-97d3-e559672eee54 req-6f887816-7694-468c-9d3c-21b30de5f0f3 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] No waiting events found dispatching network-vif-plugged-9e8240fa-3546-43eb-9cdc-4b42859690b3 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.289986] env[63293]: WARNING nova.compute.manager [req-25d535af-a68c-43b0-97d3-e559672eee54 req-6f887816-7694-468c-9d3c-21b30de5f0f3 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Received unexpected event network-vif-plugged-9e8240fa-3546-43eb-9cdc-4b42859690b3 for instance with vm_state building and task_state spawning. [ 932.425953] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5259f93b-cdb9-6c96-b9c5-fd45d121b2c5, 'name': SearchDatastore_Task, 'duration_secs': 0.090281} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.425953] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.425953] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 35cbf277-b85a-4101-beae-0125fb5a61f5/35cbf277-b85a-4101-beae-0125fb5a61f5.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 932.425953] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e31aae3b-7d76-4435-b837-3561e9cd13bd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.439410] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 932.439410] env[63293]: value = "task-1327920" [ 932.439410] env[63293]: _type = "Task" [ 932.439410] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.448742] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327920, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.451258] env[63293]: DEBUG nova.network.neutron [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Successfully updated port: 9e8240fa-3546-43eb-9cdc-4b42859690b3 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.457694] env[63293]: DEBUG oslo_vmware.api [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327919, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213281} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.458470] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.458868] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 932.460293] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 932.461147] env[63293]: INFO nova.compute.manager [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Took 0.63 seconds to destroy the instance on the hypervisor. [ 932.461147] env[63293]: DEBUG oslo.service.loopingcall [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.461274] env[63293]: DEBUG nova.compute.manager [-] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 932.461274] env[63293]: DEBUG nova.network.neutron [-] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 932.511446] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327917, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.538800] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327916, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.646834] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327914, 'name': CreateVM_Task, 'duration_secs': 1.547662} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.647072] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 932.647830] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.648028] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "[datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.648446] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 932.648720] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8dd1838-2580-4cf1-bb39-6516bb969709 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.654106] env[63293]: DEBUG nova.compute.utils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 932.659823] env[63293]: DEBUG nova.compute.manager [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 932.660242] env[63293]: DEBUG nova.network.neutron [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 932.662956] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 932.662956] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523ba521-38d7-33d7-20db-c4e0cbf0ab40" [ 932.662956] env[63293]: _type = "Task" [ 932.662956] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.674369] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "[datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.674550] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Processing image dcac372d-af95-4395-b80b-840bcd2b2ed4 {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.674858] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4/dcac372d-af95-4395-b80b-840bcd2b2ed4.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.675134] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "[datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4/dcac372d-af95-4395-b80b-840bcd2b2ed4.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.675192] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.675797] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ed3cf3a-f8df-49a7-bdf6-a22fd637b5df {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.700195] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.700484] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 932.701401] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98629274-d70d-482e-b0dc-360580eb66db {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.710207] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 932.710207] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b7f8ec-76d7-6a18-7e36-1ea9c3bc7aec" [ 932.710207] env[63293]: _type = "Task" [ 932.710207] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.726699] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Preparing fetch location {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 932.727108] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Fetch image to [datastore1] OSTACK_IMG_ab0c685b-f577-490b-900f-b8f7bc001605/OSTACK_IMG_ab0c685b-f577-490b-900f-b8f7bc001605.vmdk {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 932.727389] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Downloading stream optimized image dcac372d-af95-4395-b80b-840bcd2b2ed4 to [datastore1] OSTACK_IMG_ab0c685b-f577-490b-900f-b8f7bc001605/OSTACK_IMG_ab0c685b-f577-490b-900f-b8f7bc001605.vmdk on the data store datastore1 as vApp {{(pid=63293) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 932.727613] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Downloading image file data dcac372d-af95-4395-b80b-840bcd2b2ed4 to the ESX as VM named 'OSTACK_IMG_ab0c685b-f577-490b-900f-b8f7bc001605' {{(pid=63293) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 932.743233] env[63293]: DEBUG nova.policy [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b94ec4d2961a43ca9a331a6051fa774d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7744835de0bc4e30b1b29ee5a439cd0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.826589] env[63293]: DEBUG oslo_vmware.rw_handles [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 932.826589] env[63293]: value = "resgroup-9" [ 932.826589] env[63293]: _type = "ResourcePool" [ 932.826589] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 932.827207] env[63293]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-2bff1374-dd3b-41db-8c13-3f4f4941efe7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.854378] env[63293]: DEBUG oslo_vmware.rw_handles [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lease: (returnval){ [ 932.854378] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529c19f6-9f3a-c3a9-673f-fd3aef9a120a" [ 932.854378] env[63293]: _type = "HttpNfcLease" [ 932.854378] env[63293]: } obtained for vApp import into resource pool (val){ [ 932.854378] env[63293]: value = "resgroup-9" [ 932.854378] env[63293]: _type = "ResourcePool" [ 932.854378] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 932.854676] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the lease: (returnval){ [ 932.854676] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529c19f6-9f3a-c3a9-673f-fd3aef9a120a" [ 932.854676] env[63293]: _type = "HttpNfcLease" [ 932.854676] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 932.868017] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 932.868017] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529c19f6-9f3a-c3a9-673f-fd3aef9a120a" [ 932.868017] env[63293]: _type = "HttpNfcLease" [ 932.868017] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 932.939121] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d04240-10cf-4f2b-b7c4-28f56cbb1640 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.953220] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327920, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.956172] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c0c707-602c-4ac6-8447-d154545889ee {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.962606] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "refresh_cache-2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.962606] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "refresh_cache-2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.962606] env[63293]: DEBUG nova.network.neutron [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 933.023012] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e53b843c-7126-49ce-bcfe-960d30b03dba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.023012] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327917, 'name': PowerOnVM_Task} progress is 76%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.023012] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ed44b7-9622-4443-b429-cc429325ea32 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.043689] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327916, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.057783] env[63293]: DEBUG nova.compute.provider_tree [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.139891] env[63293]: DEBUG nova.network.neutron [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Successfully created port: c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.162025] env[63293]: DEBUG nova.compute.manager [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.291596] env[63293]: DEBUG nova.network.neutron [-] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.366915] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 933.366915] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529c19f6-9f3a-c3a9-673f-fd3aef9a120a" [ 933.366915] env[63293]: _type = "HttpNfcLease" [ 933.366915] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 933.456019] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327920, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.510191] env[63293]: DEBUG nova.network.neutron [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 933.518226] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327917, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.544478] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327916, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.564334] env[63293]: DEBUG nova.scheduler.client.report [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.777138] env[63293]: DEBUG nova.network.neutron [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Updating instance_info_cache with network_info: [{"id": "9e8240fa-3546-43eb-9cdc-4b42859690b3", "address": "fa:16:3e:04:3a:e7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e8240fa-35", "ovs_interfaceid": "9e8240fa-3546-43eb-9cdc-4b42859690b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.794541] env[63293]: INFO nova.compute.manager [-] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Took 1.33 seconds to deallocate network for instance. [ 933.869083] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 933.869083] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529c19f6-9f3a-c3a9-673f-fd3aef9a120a" [ 933.869083] env[63293]: _type = "HttpNfcLease" [ 933.869083] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 933.869512] env[63293]: DEBUG oslo_vmware.rw_handles [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 933.869512] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529c19f6-9f3a-c3a9-673f-fd3aef9a120a" [ 933.869512] env[63293]: _type = "HttpNfcLease" [ 933.869512] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 933.870522] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff2d400-df26-4e03-836d-bf9fc9fe1260 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.879999] env[63293]: DEBUG oslo_vmware.rw_handles [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d99d29-a03d-4bb3-2bdb-12d93e278a80/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 933.880925] env[63293]: DEBUG oslo_vmware.rw_handles [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d99d29-a03d-4bb3-2bdb-12d93e278a80/disk-0.vmdk. {{(pid=63293) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 933.953715] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b39d3352-0345-4d3a-8f22-33c983431e07 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.967231] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327920, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.012936] env[63293]: DEBUG oslo_vmware.api [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327917, 'name': PowerOnVM_Task, 'duration_secs': 2.344504} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.013338] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.013610] env[63293]: INFO nova.compute.manager [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Took 11.99 seconds to spawn the instance on the hypervisor. [ 934.013813] env[63293]: DEBUG nova.compute.manager [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.014714] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e33623d-89f1-4ef5-a9f4-d770cf32ba86 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.041345] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327916, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.580584} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.041639] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b176c14-05dc-43c1-a475-57283dc96350/8b176c14-05dc-43c1-a475-57283dc96350.vmdk to [datastore2] 2557bded-6ccb-4c0f-8b97-34529b4bb740/2557bded-6ccb-4c0f-8b97-34529b4bb740.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.042462] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836b5b08-98bf-4bfe-ab6a-e05d8ebf619d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.065963] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 2557bded-6ccb-4c0f-8b97-34529b4bb740/2557bded-6ccb-4c0f-8b97-34529b4bb740.vmdk or device None with type streamOptimized {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.067024] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-133de838-a32f-408c-9b87-de627dec9c66 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.092831] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 934.092831] env[63293]: value = "task-1327922" [ 934.092831] env[63293]: _type = "Task" [ 934.092831] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.101134] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327922, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.171492] env[63293]: DEBUG nova.compute.manager [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.277754] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "refresh_cache-2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.277754] env[63293]: DEBUG nova.compute.manager [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Instance network_info: |[{"id": "9e8240fa-3546-43eb-9cdc-4b42859690b3", "address": "fa:16:3e:04:3a:e7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e8240fa-35", "ovs_interfaceid": "9e8240fa-3546-43eb-9cdc-4b42859690b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 934.277754] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:3a:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91c1da19-ab68-4127-bacd-accbaff19651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9e8240fa-3546-43eb-9cdc-4b42859690b3', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 934.286379] env[63293]: DEBUG oslo.service.loopingcall [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.286725] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 934.286989] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ab056c3-834d-4202-8615-919a071b4016 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.303285] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.310044] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.310044] env[63293]: value = "task-1327923" [ 934.310044] env[63293]: _type = "Task" [ 934.310044] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.318616] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327923, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.321432] env[63293]: DEBUG nova.compute.manager [req-07a8f3ec-3194-41c8-9c2f-803b94f3ddb3 req-1c6395da-5f0d-47c2-8e37-089bd82f5be4 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Received event network-changed-9e8240fa-3546-43eb-9cdc-4b42859690b3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.321648] env[63293]: DEBUG nova.compute.manager [req-07a8f3ec-3194-41c8-9c2f-803b94f3ddb3 req-1c6395da-5f0d-47c2-8e37-089bd82f5be4 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Refreshing instance network info cache due to event network-changed-9e8240fa-3546-43eb-9cdc-4b42859690b3. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.321921] env[63293]: DEBUG oslo_concurrency.lockutils [req-07a8f3ec-3194-41c8-9c2f-803b94f3ddb3 req-1c6395da-5f0d-47c2-8e37-089bd82f5be4 service nova] Acquiring lock "refresh_cache-2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.322089] env[63293]: DEBUG oslo_concurrency.lockutils [req-07a8f3ec-3194-41c8-9c2f-803b94f3ddb3 req-1c6395da-5f0d-47c2-8e37-089bd82f5be4 service nova] Acquired lock "refresh_cache-2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.322285] env[63293]: DEBUG nova.network.neutron [req-07a8f3ec-3194-41c8-9c2f-803b94f3ddb3 req-1c6395da-5f0d-47c2-8e37-089bd82f5be4 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Refreshing network info cache for port 9e8240fa-3546-43eb-9cdc-4b42859690b3 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 934.459037] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327920, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.651055} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.460528] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 35cbf277-b85a-4101-beae-0125fb5a61f5/35cbf277-b85a-4101-beae-0125fb5a61f5.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.460762] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.461080] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-010c12e8-679a-45b2-a9f5-44b1073c6140 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.468974] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 934.468974] env[63293]: value = "task-1327924" [ 934.468974] env[63293]: _type = "Task" [ 934.468974] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.479742] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327924, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.535321] env[63293]: INFO nova.compute.manager [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Took 22.35 seconds to build instance. [ 934.589749] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.439s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.595032] env[63293]: DEBUG oslo_concurrency.lockutils [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.802s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.595271] env[63293]: DEBUG nova.objects.instance [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lazy-loading 'resources' on Instance uuid 5b1264b4-2554-4c62-9ff9-4616b03d3609 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.606978] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327922, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.821845] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327923, 'name': CreateVM_Task, 'duration_secs': 0.402} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.822040] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 934.822878] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.823195] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.823725] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.826462] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1f389cb-befc-487d-8f38-87e92fdbb125 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.836163] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 934.836163] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f31f95-6da9-25e4-6daf-8985f561ecaa" [ 934.836163] env[63293]: _type = "Task" [ 934.836163] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.849249] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f31f95-6da9-25e4-6daf-8985f561ecaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.981499] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327924, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090341} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.983332] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 934.984120] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01199b18-8b08-4a41-8be6-d26adc89e92f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.007612] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 35cbf277-b85a-4101-beae-0125fb5a61f5/35cbf277-b85a-4101-beae-0125fb5a61f5.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.013086] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be879f75-dc02-4145-875a-3ce4c083db25 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.048638] env[63293]: DEBUG oslo_concurrency.lockutils [None req-3a4a7af4-dc5e-41a3-ad72-a5ff0ecda818 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "f429869f-9e9f-443e-a980-67a80c3bd799" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.884s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.053824] env[63293]: DEBUG nova.network.neutron [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Successfully updated port: c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.062981] env[63293]: DEBUG nova.virt.hardware [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.062981] env[63293]: DEBUG nova.virt.hardware [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.062981] env[63293]: DEBUG nova.virt.hardware [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.064014] env[63293]: DEBUG nova.virt.hardware [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.064014] env[63293]: DEBUG nova.virt.hardware [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.064014] env[63293]: DEBUG nova.virt.hardware [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.064014] env[63293]: DEBUG nova.virt.hardware [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.064014] env[63293]: DEBUG nova.virt.hardware [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.064014] env[63293]: DEBUG nova.virt.hardware [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.064394] env[63293]: DEBUG nova.virt.hardware [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.064394] env[63293]: DEBUG nova.virt.hardware [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.066729] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7448e9bb-c898-45cb-bae0-95b88284f0bd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.069416] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 935.069416] env[63293]: value = "task-1327925" [ 935.069416] env[63293]: _type = "Task" [ 935.069416] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.082041] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a40e3c-69ca-423e-91e6-689f09930784 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.091569] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327925, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.125047] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327922, 'name': ReconfigVM_Task, 'duration_secs': 0.568089} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.125253] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 2557bded-6ccb-4c0f-8b97-34529b4bb740/2557bded-6ccb-4c0f-8b97-34529b4bb740.vmdk or device None with type streamOptimized {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.125802] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2eb65cce-a37d-44be-80bf-fa5fe94e9b19 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.136051] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 935.136051] env[63293]: value = "task-1327926" [ 935.136051] env[63293]: _type = "Task" [ 935.136051] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.147940] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327926, 'name': Rename_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.178609] env[63293]: INFO nova.scheduler.client.report [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted allocation for migration 5484fbb9-cb45-4302-9128-7eba0d080fc7 [ 935.373894] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f31f95-6da9-25e4-6daf-8985f561ecaa, 'name': SearchDatastore_Task, 'duration_secs': 0.012864} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.381440] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.381787] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.382119] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.382458] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.382666] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.384024] env[63293]: DEBUG nova.compute.manager [req-0e8a21f1-d6ba-4af9-9118-5ba7a12f12a8 req-0e094ef4-f561-4a53-871c-c6613710e5b6 service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Received event network-vif-plugged-c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.384286] env[63293]: DEBUG oslo_concurrency.lockutils [req-0e8a21f1-d6ba-4af9-9118-5ba7a12f12a8 req-0e094ef4-f561-4a53-871c-c6613710e5b6 service nova] Acquiring lock "828180c0-b7f8-4666-9f5d-e6741ef6495a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.385013] env[63293]: DEBUG oslo_concurrency.lockutils [req-0e8a21f1-d6ba-4af9-9118-5ba7a12f12a8 req-0e094ef4-f561-4a53-871c-c6613710e5b6 service nova] Lock "828180c0-b7f8-4666-9f5d-e6741ef6495a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.385013] env[63293]: DEBUG oslo_concurrency.lockutils [req-0e8a21f1-d6ba-4af9-9118-5ba7a12f12a8 req-0e094ef4-f561-4a53-871c-c6613710e5b6 service nova] Lock "828180c0-b7f8-4666-9f5d-e6741ef6495a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.385013] env[63293]: DEBUG nova.compute.manager [req-0e8a21f1-d6ba-4af9-9118-5ba7a12f12a8 req-0e094ef4-f561-4a53-871c-c6613710e5b6 service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] No waiting events found dispatching network-vif-plugged-c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.385590] env[63293]: WARNING nova.compute.manager [req-0e8a21f1-d6ba-4af9-9118-5ba7a12f12a8 req-0e094ef4-f561-4a53-871c-c6613710e5b6 service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Received unexpected event network-vif-plugged-c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242 for instance with vm_state building and task_state spawning. [ 935.388849] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8fd3cb89-2142-4794-a071-c8c40307481d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.399282] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.399550] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 935.401203] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99a40e01-6272-4652-8d31-5131e99d02a1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.413810] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 935.413810] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529fa835-36d8-c9ce-42c0-c9f246d163e0" [ 935.413810] env[63293]: _type = "Task" [ 935.413810] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.426064] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529fa835-36d8-c9ce-42c0-c9f246d163e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.440377] env[63293]: DEBUG nova.network.neutron [req-07a8f3ec-3194-41c8-9c2f-803b94f3ddb3 req-1c6395da-5f0d-47c2-8e37-089bd82f5be4 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Updated VIF entry in instance network info cache for port 9e8240fa-3546-43eb-9cdc-4b42859690b3. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 935.440763] env[63293]: DEBUG nova.network.neutron [req-07a8f3ec-3194-41c8-9c2f-803b94f3ddb3 req-1c6395da-5f0d-47c2-8e37-089bd82f5be4 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Updating instance_info_cache with network_info: [{"id": "9e8240fa-3546-43eb-9cdc-4b42859690b3", "address": "fa:16:3e:04:3a:e7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e8240fa-35", "ovs_interfaceid": "9e8240fa-3546-43eb-9cdc-4b42859690b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.446251] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ce9d5d-5933-4e92-ac05-8ec23cda9eac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.457434] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d73ea9-06f4-4569-81c4-948fc4b4aeb5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.511024] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a787072-40bd-42d1-8749-6618a93de8cb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.520671] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9485c7a6-b801-43b2-b293-3fa3ca60493e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.537800] env[63293]: DEBUG nova.compute.provider_tree [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.562179] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "refresh_cache-828180c0-b7f8-4666-9f5d-e6741ef6495a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.562179] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "refresh_cache-828180c0-b7f8-4666-9f5d-e6741ef6495a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.562179] env[63293]: DEBUG nova.network.neutron [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.582853] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327925, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.648775] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327926, 'name': Rename_Task, 'duration_secs': 0.285909} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.652624] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 935.652972] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49b8d97f-2559-43f5-a91c-28182a23621a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.661072] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 935.661072] env[63293]: value = "task-1327927" [ 935.661072] env[63293]: _type = "Task" [ 935.661072] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.674224] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327927, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.689316] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9841daad-4037-49cb-8128-e42f99b01ad7 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 13.159s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.749617] env[63293]: DEBUG oslo_vmware.rw_handles [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Completed reading data from the image iterator. {{(pid=63293) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 935.751292] env[63293]: DEBUG oslo_vmware.rw_handles [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d99d29-a03d-4bb3-2bdb-12d93e278a80/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 935.751292] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebab581f-39fa-4b48-8ff5-6955a906fb3e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.759912] env[63293]: DEBUG oslo_vmware.rw_handles [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d99d29-a03d-4bb3-2bdb-12d93e278a80/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 935.760061] env[63293]: DEBUG oslo_vmware.rw_handles [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d99d29-a03d-4bb3-2bdb-12d93e278a80/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 935.760422] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ad950282-1200-441e-b492-da33048bbd3b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.924155] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529fa835-36d8-c9ce-42c0-c9f246d163e0, 'name': SearchDatastore_Task, 'duration_secs': 0.010359} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.924972] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a1eb4d4-8a79-4210-b5b3-319b1acc541b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.930818] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 935.930818] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52da5179-8772-80bf-adf0-3444df826322" [ 935.930818] env[63293]: _type = "Task" [ 935.930818] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.939353] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52da5179-8772-80bf-adf0-3444df826322, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.944212] env[63293]: DEBUG oslo_concurrency.lockutils [req-07a8f3ec-3194-41c8-9c2f-803b94f3ddb3 req-1c6395da-5f0d-47c2-8e37-089bd82f5be4 service nova] Releasing lock "refresh_cache-2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.944517] env[63293]: DEBUG nova.compute.manager [req-07a8f3ec-3194-41c8-9c2f-803b94f3ddb3 req-1c6395da-5f0d-47c2-8e37-089bd82f5be4 service nova] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Received event network-vif-deleted-94365438-f2bc-4062-add0-5c8519bf1a6d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.956686] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.956999] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.043791] env[63293]: DEBUG nova.scheduler.client.report [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.084503] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327925, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.098406] env[63293]: DEBUG oslo_vmware.rw_handles [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d99d29-a03d-4bb3-2bdb-12d93e278a80/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 936.098699] env[63293]: INFO nova.virt.vmwareapi.images [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Downloaded image file data dcac372d-af95-4395-b80b-840bcd2b2ed4 [ 936.099731] env[63293]: DEBUG nova.network.neutron [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.102544] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d2ee27-a332-487f-b36c-960f6973d143 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.120879] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b020dcf-931b-4a84-93d7-7b7a23fca2f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.169079] env[63293]: INFO nova.virt.vmwareapi.images [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] The imported VM was unregistered [ 936.171184] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Caching image {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 936.171457] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating directory with path [datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4 {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.175134] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c04d318d-06c7-41b6-a8f4-4830364b1b9d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.177656] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327927, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.191814] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Created directory with path [datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4 {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.192065] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_ab0c685b-f577-490b-900f-b8f7bc001605/OSTACK_IMG_ab0c685b-f577-490b-900f-b8f7bc001605.vmdk to [datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4/dcac372d-af95-4395-b80b-840bcd2b2ed4.vmdk. {{(pid=63293) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 936.192700] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e6838ccb-cda3-4e11-a3da-ec3abcdacd4e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.199613] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 936.199613] env[63293]: value = "task-1327929" [ 936.199613] env[63293]: _type = "Task" [ 936.199613] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.208362] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327929, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.255096] env[63293]: DEBUG nova.network.neutron [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Updating instance_info_cache with network_info: [{"id": "c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242", "address": "fa:16:3e:63:cc:f8", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1fa4ad6-2d", "ovs_interfaceid": "c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.444651] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52da5179-8772-80bf-adf0-3444df826322, 'name': SearchDatastore_Task, 'duration_secs': 0.022049} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.445030] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.445349] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535/2f55ac2a-63d1-4713-ab0c-47fb6cf7d535.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 936.447520] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-062fdefc-f795-4cc5-ad1b-79dddc9fcd06 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.455838] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 936.455838] env[63293]: value = "task-1327930" [ 936.455838] env[63293]: _type = "Task" [ 936.455838] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.460324] env[63293]: DEBUG nova.compute.manager [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 936.469144] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327930, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.549034] env[63293]: DEBUG oslo_concurrency.lockutils [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.954s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.554025] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.573s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.554025] env[63293]: DEBUG nova.objects.instance [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Lazy-loading 'resources' on Instance uuid 8625a358-ec48-46b9-8738-c49a2ba58362 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.580537] env[63293]: INFO nova.scheduler.client.report [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Deleted allocations for instance 5b1264b4-2554-4c62-9ff9-4616b03d3609 [ 936.590582] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327925, 'name': ReconfigVM_Task, 'duration_secs': 1.390637} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.591057] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 35cbf277-b85a-4101-beae-0125fb5a61f5/35cbf277-b85a-4101-beae-0125fb5a61f5.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.591943] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94667050-3b7b-4578-b741-2215afd0c162 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.601074] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 936.601074] env[63293]: value = "task-1327931" [ 936.601074] env[63293]: _type = "Task" [ 936.601074] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.609945] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327931, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.676509] env[63293]: DEBUG oslo_vmware.api [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327927, 'name': PowerOnVM_Task, 'duration_secs': 0.648519} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.677395] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 936.677395] env[63293]: INFO nova.compute.manager [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Took 17.07 seconds to spawn the instance on the hypervisor. [ 936.677732] env[63293]: DEBUG nova.compute.manager [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.678686] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71bae73a-09c3-4079-b60e-4962d5b548cf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.714280] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327929, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.758687] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "refresh_cache-828180c0-b7f8-4666-9f5d-e6741ef6495a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.759055] env[63293]: DEBUG nova.compute.manager [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Instance network_info: |[{"id": "c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242", "address": "fa:16:3e:63:cc:f8", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1fa4ad6-2d", "ovs_interfaceid": "c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 936.759552] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:cc:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ef746c57-cd18-4883-a0e9-c52937aaf41d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.767401] env[63293]: DEBUG oslo.service.loopingcall [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.768095] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 936.768383] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a8748ca6-56f8-4f6e-b7f9-13bcb5d22208 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.791675] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.791675] env[63293]: value = "task-1327932" [ 936.791675] env[63293]: _type = "Task" [ 936.791675] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.799009] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327932, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.970456] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327930, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.988338] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.090016] env[63293]: DEBUG oslo_concurrency.lockutils [None req-01db8596-ad3e-4ad6-b871-7531fce79fb5 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "5b1264b4-2554-4c62-9ff9-4616b03d3609" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.860s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.113067] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327931, 'name': Rename_Task, 'duration_secs': 0.162712} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.116104] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 937.117848] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96aecdb0-3d36-497f-b4f9-2ec6b08ab219 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.128756] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 937.128756] env[63293]: value = "task-1327933" [ 937.128756] env[63293]: _type = "Task" [ 937.128756] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.147810] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327933, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.200894] env[63293]: INFO nova.compute.manager [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Took 34.24 seconds to build instance. [ 937.213079] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327929, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.302763] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327932, 'name': CreateVM_Task, 'duration_secs': 0.423071} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.304240] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 937.305107] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f6a4e3-d2ce-4504-ac59-f209f9aa7736 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.308648] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.308838] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.309242] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 937.309556] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5db1e129-2668-465e-869b-bc9be76f8147 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.313991] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.314423] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.314753] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.315040] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.315332] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.322604] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 937.322604] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d0211b-a84c-0e19-27ba-bd66c2c1bedd" [ 937.322604] env[63293]: _type = "Task" [ 937.322604] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.323382] env[63293]: INFO nova.compute.manager [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Terminating instance [ 937.327158] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb124786-f031-41d0-891e-ad5f241bb916 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.332034] env[63293]: DEBUG nova.compute.manager [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 937.332302] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 937.337342] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e831e96a-d3d4-4e1e-ba87-1b8d492cab85 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.346863] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d0211b-a84c-0e19-27ba-bd66c2c1bedd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.378872] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.379936] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-599b2072-3373-4cbb-a797-0730005130eb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.382356] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f6fb4a-0593-48f4-a3ca-baef5b5533c6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.392436] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33e4485-9dd7-484e-b4af-ff7f7d70dc63 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.398475] env[63293]: DEBUG oslo_vmware.api [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 937.398475] env[63293]: value = "task-1327934" [ 937.398475] env[63293]: _type = "Task" [ 937.398475] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.414051] env[63293]: DEBUG nova.compute.provider_tree [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.420100] env[63293]: DEBUG oslo_vmware.api [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327934, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.459894] env[63293]: DEBUG nova.compute.manager [req-c3fe406f-d3d1-4510-95a8-7d8b1f6f84fb req-c3fe0ec7-fc0b-40d0-aa1a-698f0f10461d service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Received event network-changed-c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.460250] env[63293]: DEBUG nova.compute.manager [req-c3fe406f-d3d1-4510-95a8-7d8b1f6f84fb req-c3fe0ec7-fc0b-40d0-aa1a-698f0f10461d service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Refreshing instance network info cache due to event network-changed-c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.460540] env[63293]: DEBUG oslo_concurrency.lockutils [req-c3fe406f-d3d1-4510-95a8-7d8b1f6f84fb req-c3fe0ec7-fc0b-40d0-aa1a-698f0f10461d service nova] Acquiring lock "refresh_cache-828180c0-b7f8-4666-9f5d-e6741ef6495a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.460892] env[63293]: DEBUG oslo_concurrency.lockutils [req-c3fe406f-d3d1-4510-95a8-7d8b1f6f84fb req-c3fe0ec7-fc0b-40d0-aa1a-698f0f10461d service nova] Acquired lock "refresh_cache-828180c0-b7f8-4666-9f5d-e6741ef6495a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.461074] env[63293]: DEBUG nova.network.neutron [req-c3fe406f-d3d1-4510-95a8-7d8b1f6f84fb req-c3fe0ec7-fc0b-40d0-aa1a-698f0f10461d service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Refreshing network info cache for port c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.473925] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327930, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.545912] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "2557bded-6ccb-4c0f-8b97-34529b4bb740" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.625173] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.625541] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.625730] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.625976] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.626213] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.628687] env[63293]: INFO nova.compute.manager [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Terminating instance [ 937.631059] env[63293]: DEBUG nova.compute.manager [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 937.631283] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 937.632280] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deffafc1-a9f6-4567-ae4e-ef7444d4ac72 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.648155] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327933, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.651695] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.652120] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea5176ec-daa2-43fb-a645-b8b545b1d72a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.660768] env[63293]: DEBUG oslo_vmware.api [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 937.660768] env[63293]: value = "task-1327935" [ 937.660768] env[63293]: _type = "Task" [ 937.660768] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.671087] env[63293]: DEBUG oslo_vmware.api [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327935, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.708065] env[63293]: DEBUG oslo_concurrency.lockutils [None req-35ca58c9-74df-4c45-9a29-f4a18e9dd4af tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "2557bded-6ccb-4c0f-8b97-34529b4bb740" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.763s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.708426] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "2557bded-6ccb-4c0f-8b97-34529b4bb740" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.163s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.708808] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "2557bded-6ccb-4c0f-8b97-34529b4bb740-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.709123] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "2557bded-6ccb-4c0f-8b97-34529b4bb740-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.709373] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "2557bded-6ccb-4c0f-8b97-34529b4bb740-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.712492] env[63293]: INFO nova.compute.manager [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Terminating instance [ 937.720084] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327929, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.720726] env[63293]: DEBUG nova.compute.manager [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 937.721045] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 937.722196] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed63f77e-f12d-48d5-b87c-82cccaf8828b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.731958] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.732503] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70c89820-8bc8-47f5-8813-92cae9fddcd7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.741563] env[63293]: DEBUG oslo_vmware.api [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 937.741563] env[63293]: value = "task-1327936" [ 937.741563] env[63293]: _type = "Task" [ 937.741563] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.750742] env[63293]: DEBUG oslo_vmware.api [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327936, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.844334] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d0211b-a84c-0e19-27ba-bd66c2c1bedd, 'name': SearchDatastore_Task, 'duration_secs': 0.087475} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.844908] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.845358] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 937.845859] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.846067] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.846395] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.846867] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e850fe75-0f8f-451b-9ac7-e00c52573321 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.865323] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.865556] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 937.866488] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d7a1a0c-04d8-4941-a47b-7c2d17a9d577 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.873567] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 937.873567] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c2d348-59b6-ca2f-cb5e-cb1e3264255c" [ 937.873567] env[63293]: _type = "Task" [ 937.873567] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.882570] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c2d348-59b6-ca2f-cb5e-cb1e3264255c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.912419] env[63293]: DEBUG oslo_vmware.api [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327934, 'name': PowerOffVM_Task, 'duration_secs': 0.248655} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.912924] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 937.913059] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 937.913315] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6dfd179-1da5-4b1d-91d3-80030cfb93c9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.916182] env[63293]: DEBUG nova.scheduler.client.report [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.948141] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.948427] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.971740] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327930, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.993764] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 937.993764] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 937.993764] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleting the datastore file [datastore2] 76ebe2ec-332c-4d3a-957e-d0c928dafdc7 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 937.993764] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8ec51520-db45-4437-9ace-a71437c9c6f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.005687] env[63293]: DEBUG oslo_vmware.api [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 938.005687] env[63293]: value = "task-1327938" [ 938.005687] env[63293]: _type = "Task" [ 938.005687] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.015380] env[63293]: DEBUG oslo_vmware.api [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327938, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.147649] env[63293]: DEBUG oslo_vmware.api [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327933, 'name': PowerOnVM_Task, 'duration_secs': 0.557229} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.147950] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 938.148204] env[63293]: INFO nova.compute.manager [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Took 11.76 seconds to spawn the instance on the hypervisor. [ 938.148417] env[63293]: DEBUG nova.compute.manager [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.149309] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042bf6a9-159f-4ec4-8b50-0afda5f81052 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.171151] env[63293]: DEBUG oslo_vmware.api [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327935, 'name': PowerOffVM_Task, 'duration_secs': 0.244735} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.171578] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.171683] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.171916] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ceb8e75f-7da4-4864-8cd3-5ab173da3eaa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.215081] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327929, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.249638] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.250013] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.250332] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Deleting the datastore file [datastore2] eec9bb88-c5c4-4d8a-8a30-84b8d89e153d {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.251710] env[63293]: DEBUG nova.network.neutron [req-c3fe406f-d3d1-4510-95a8-7d8b1f6f84fb req-c3fe0ec7-fc0b-40d0-aa1a-698f0f10461d service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Updated VIF entry in instance network info cache for port c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 938.252245] env[63293]: DEBUG nova.network.neutron [req-c3fe406f-d3d1-4510-95a8-7d8b1f6f84fb req-c3fe0ec7-fc0b-40d0-aa1a-698f0f10461d service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Updating instance_info_cache with network_info: [{"id": "c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242", "address": "fa:16:3e:63:cc:f8", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1fa4ad6-2d", "ovs_interfaceid": "c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.254555] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c57273f8-3fee-45dc-b540-94e9a9738885 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.263767] env[63293]: DEBUG oslo_vmware.api [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327936, 'name': PowerOffVM_Task, 'duration_secs': 0.236905} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.265344] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.265705] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.266854] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-110326d0-465e-492c-a46e-886c79b0b573 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.272772] env[63293]: DEBUG oslo_vmware.api [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for the task: (returnval){ [ 938.272772] env[63293]: value = "task-1327940" [ 938.272772] env[63293]: _type = "Task" [ 938.272772] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.288142] env[63293]: DEBUG oslo_vmware.api [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327940, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.381619] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.382121] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.382456] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleting the datastore file [datastore2] 2557bded-6ccb-4c0f-8b97-34529b4bb740 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.383435] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e880fed-67df-45b4-8478-40a23e101cf7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.392807] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c2d348-59b6-ca2f-cb5e-cb1e3264255c, 'name': SearchDatastore_Task, 'duration_secs': 0.082059} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.394730] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-719c95ee-de27-4a01-bc2a-0a609cfec7a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.400306] env[63293]: DEBUG oslo_vmware.api [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 938.400306] env[63293]: value = "task-1327942" [ 938.400306] env[63293]: _type = "Task" [ 938.400306] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.407751] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 938.407751] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b0b6c9-4aec-a340-4e69-a529ed7de437" [ 938.407751] env[63293]: _type = "Task" [ 938.407751] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.419093] env[63293]: DEBUG oslo_vmware.api [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327942, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.427235] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.875s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.430814] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b0b6c9-4aec-a340-4e69-a529ed7de437, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.431988] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.128s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.431988] env[63293]: DEBUG nova.objects.instance [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lazy-loading 'resources' on Instance uuid fe55a9b7-28b5-4b04-84f1-6b85b579a144 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.450611] env[63293]: DEBUG nova.compute.manager [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 938.468547] env[63293]: INFO nova.scheduler.client.report [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Deleted allocations for instance 8625a358-ec48-46b9-8738-c49a2ba58362 [ 938.478820] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327930, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.514521] env[63293]: DEBUG oslo_vmware.api [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327938, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.371101} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.514837] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.515044] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 938.515235] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 938.515415] env[63293]: INFO nova.compute.manager [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Took 1.18 seconds to destroy the instance on the hypervisor. [ 938.515673] env[63293]: DEBUG oslo.service.loopingcall [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.515912] env[63293]: DEBUG nova.compute.manager [-] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 938.516030] env[63293]: DEBUG nova.network.neutron [-] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 938.671081] env[63293]: INFO nova.compute.manager [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Took 21.47 seconds to build instance. [ 938.714737] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327929, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.759114] env[63293]: DEBUG oslo_concurrency.lockutils [req-c3fe406f-d3d1-4510-95a8-7d8b1f6f84fb req-c3fe0ec7-fc0b-40d0-aa1a-698f0f10461d service nova] Releasing lock "refresh_cache-828180c0-b7f8-4666-9f5d-e6741ef6495a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.785285] env[63293]: DEBUG oslo_vmware.api [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Task: {'id': task-1327940, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.398412} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.785659] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.785801] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 938.786206] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 938.786206] env[63293]: INFO nova.compute.manager [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 938.786435] env[63293]: DEBUG oslo.service.loopingcall [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.786789] env[63293]: DEBUG nova.compute.manager [-] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 938.786789] env[63293]: DEBUG nova.network.neutron [-] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 938.911428] env[63293]: DEBUG oslo_vmware.api [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327942, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.405669} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.914170] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.914399] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 938.914580] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 938.914764] env[63293]: INFO nova.compute.manager [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Took 1.19 seconds to destroy the instance on the hypervisor. [ 938.915088] env[63293]: DEBUG oslo.service.loopingcall [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.915644] env[63293]: DEBUG nova.compute.manager [-] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 938.915745] env[63293]: DEBUG nova.network.neutron [-] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 938.923050] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b0b6c9-4aec-a340-4e69-a529ed7de437, 'name': SearchDatastore_Task, 'duration_secs': 0.123365} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.924959] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.925163] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 828180c0-b7f8-4666-9f5d-e6741ef6495a/828180c0-b7f8-4666-9f5d-e6741ef6495a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 938.926376] env[63293]: DEBUG nova.compute.manager [req-279f0ecf-cf62-40e5-b35f-d3c6ae61f5e5 req-6511e354-cc28-4f6e-aeaf-561da5ced456 service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Received event network-vif-deleted-46d5563f-41a2-48ab-9d32-3aa394a13da8 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.926564] env[63293]: INFO nova.compute.manager [req-279f0ecf-cf62-40e5-b35f-d3c6ae61f5e5 req-6511e354-cc28-4f6e-aeaf-561da5ced456 service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Neutron deleted interface 46d5563f-41a2-48ab-9d32-3aa394a13da8; detaching it from the instance and deleting it from the info cache [ 938.926762] env[63293]: DEBUG nova.network.neutron [req-279f0ecf-cf62-40e5-b35f-d3c6ae61f5e5 req-6511e354-cc28-4f6e-aeaf-561da5ced456 service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.929239] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8881c6aa-86b0-4252-b8cc-a138c1a3c911 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.939241] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 938.939241] env[63293]: value = "task-1327943" [ 938.939241] env[63293]: _type = "Task" [ 938.939241] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.950313] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.973140] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327930, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.429435} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.973140] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535/2f55ac2a-63d1-4713-ab0c-47fb6cf7d535.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 938.973140] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.976028] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55186510-1d0d-40fc-8c2b-80ab605d89be {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.981702] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.984849] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9437c235-23dd-49c3-8c8a-6e424ef5a0d3 tempest-ServerMetadataTestJSON-997880905 tempest-ServerMetadataTestJSON-997880905-project-member] Lock "8625a358-ec48-46b9-8738-c49a2ba58362" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.207s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.987924] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 938.987924] env[63293]: value = "task-1327944" [ 938.987924] env[63293]: _type = "Task" [ 938.987924] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.996894] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327944, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.173719] env[63293]: DEBUG oslo_concurrency.lockutils [None req-02bb1585-0b30-4109-8a7c-b70da258ef75 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "35cbf277-b85a-4101-beae-0125fb5a61f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.986s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.216945] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327929, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.852635} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.217139] env[63293]: INFO nova.virt.vmwareapi.ds_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_ab0c685b-f577-490b-900f-b8f7bc001605/OSTACK_IMG_ab0c685b-f577-490b-900f-b8f7bc001605.vmdk to [datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4/dcac372d-af95-4395-b80b-840bcd2b2ed4.vmdk. [ 939.217297] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Cleaning up location [datastore1] OSTACK_IMG_ab0c685b-f577-490b-900f-b8f7bc001605 {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 939.219110] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_ab0c685b-f577-490b-900f-b8f7bc001605 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.219110] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c28fcb43-c825-4129-8adf-5726b3085c6a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.224127] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf873ec-7eae-4e18-b300-b58d858346dc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.228609] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 939.228609] env[63293]: value = "task-1327945" [ 939.228609] env[63293]: _type = "Task" [ 939.228609] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.235674] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4382ef22-2ef3-449f-98b9-78afdf68463c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.242785] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327945, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.275220] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d931e8-4f3a-4fb7-840c-cf6cad37a264 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.283112] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9053f643-3c27-4761-8aad-fb12c4a4b501 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.297702] env[63293]: DEBUG nova.compute.provider_tree [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.389526] env[63293]: DEBUG nova.network.neutron [-] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.429454] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c976a3b-753b-4f86-aa60-ca56283d964e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.439288] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a1580a-36af-4be6-9177-46fc2fb65de9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.461815] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.475263] env[63293]: DEBUG nova.compute.manager [req-279f0ecf-cf62-40e5-b35f-d3c6ae61f5e5 req-6511e354-cc28-4f6e-aeaf-561da5ced456 service nova] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Detach interface failed, port_id=46d5563f-41a2-48ab-9d32-3aa394a13da8, reason: Instance 76ebe2ec-332c-4d3a-957e-d0c928dafdc7 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 939.489931] env[63293]: DEBUG nova.compute.manager [req-a01d65de-657a-4c34-8805-4b720554b028 req-14312dfe-680c-4ced-ad86-536280e80244 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Received event network-vif-deleted-dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.490151] env[63293]: INFO nova.compute.manager [req-a01d65de-657a-4c34-8805-4b720554b028 req-14312dfe-680c-4ced-ad86-536280e80244 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Neutron deleted interface dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77; detaching it from the instance and deleting it from the info cache [ 939.490337] env[63293]: DEBUG nova.network.neutron [req-a01d65de-657a-4c34-8805-4b720554b028 req-14312dfe-680c-4ced-ad86-536280e80244 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.504649] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327944, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.259883} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.504649] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.504649] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48672e7-0be3-4be0-8c1b-595b95a34c3a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.529025] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535/2f55ac2a-63d1-4713-ab0c-47fb6cf7d535.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.529757] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-adbc8be7-efd6-4980-a103-fd59560583a4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.550620] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 939.550620] env[63293]: value = "task-1327946" [ 939.550620] env[63293]: _type = "Task" [ 939.550620] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.558972] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327946, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.616974] env[63293]: DEBUG nova.network.neutron [-] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.744082] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327945, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16373} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.744082] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.744082] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "[datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4/dcac372d-af95-4395-b80b-840bcd2b2ed4.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.744082] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4/dcac372d-af95-4395-b80b-840bcd2b2ed4.vmdk to [datastore1] ad585ebb-2072-45df-b645-94c9fa93576b/ad585ebb-2072-45df-b645-94c9fa93576b.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 939.744082] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8cc2488-3dea-4079-a028-6f3c7fefbb26 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.752673] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 939.752673] env[63293]: value = "task-1327947" [ 939.752673] env[63293]: _type = "Task" [ 939.752673] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.762239] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327947, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.802559] env[63293]: DEBUG nova.scheduler.client.report [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.892489] env[63293]: INFO nova.compute.manager [-] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Took 1.38 seconds to deallocate network for instance. [ 939.893236] env[63293]: DEBUG nova.network.neutron [-] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.967031] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327943, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.999144] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92f5e1df-b4f8-4801-a2ec-3ce08840d143 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.013241] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5111a3f0-9d6f-415c-9cbd-a3ff9375e3db {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.064995] env[63293]: DEBUG nova.compute.manager [req-a01d65de-657a-4c34-8805-4b720554b028 req-14312dfe-680c-4ced-ad86-536280e80244 service nova] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Detach interface failed, port_id=dd9b3d1b-eb90-4e09-8bfd-94bb6ac13a77, reason: Instance eec9bb88-c5c4-4d8a-8a30-84b8d89e153d could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 940.086828] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327946, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.123024] env[63293]: INFO nova.compute.manager [-] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Took 1.33 seconds to deallocate network for instance. [ 940.139139] env[63293]: DEBUG oslo_vmware.rw_handles [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52161e4e-e143-4795-e91e-2a2ab4ab2a5c/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 940.139139] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c68567-c60e-470a-b6de-a878bf8f4729 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.146940] env[63293]: DEBUG oslo_vmware.rw_handles [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52161e4e-e143-4795-e91e-2a2ab4ab2a5c/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 940.147958] env[63293]: ERROR oslo_vmware.rw_handles [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52161e4e-e143-4795-e91e-2a2ab4ab2a5c/disk-0.vmdk due to incomplete transfer. [ 940.148407] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-58a5f009-fd77-4e4c-8e03-afcf5543bbb0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.163169] env[63293]: DEBUG oslo_vmware.rw_handles [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52161e4e-e143-4795-e91e-2a2ab4ab2a5c/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 940.163169] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Uploaded image 66bcd27d-468f-428b-b4cb-6fed31de06c3 to the Glance image server {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 940.165801] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Destroying the VM {{(pid=63293) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 940.166215] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0dcb3cb1-2a87-45d9-bc47-d43006969840 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.178025] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 940.178025] env[63293]: value = "task-1327948" [ 940.178025] env[63293]: _type = "Task" [ 940.178025] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.187656] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327948, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.263293] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327947, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.311248] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.879s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.313574] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.325s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.315773] env[63293]: INFO nova.compute.claims [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 940.338142] env[63293]: INFO nova.scheduler.client.report [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleted allocations for instance fe55a9b7-28b5-4b04-84f1-6b85b579a144 [ 940.397467] env[63293]: INFO nova.compute.manager [-] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Took 1.48 seconds to deallocate network for instance. [ 940.403495] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.465434] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327943, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.161627} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.465869] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 828180c0-b7f8-4666-9f5d-e6741ef6495a/828180c0-b7f8-4666-9f5d-e6741ef6495a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 940.466109] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.466805] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8907cca4-3099-4edd-93bc-a417b529dee1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.480123] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 940.480123] env[63293]: value = "task-1327949" [ 940.480123] env[63293]: _type = "Task" [ 940.480123] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.490319] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327949, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.579829] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327946, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.629700] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.688997] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327948, 'name': Destroy_Task, 'duration_secs': 0.447331} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.689327] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Destroyed the VM [ 940.689599] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Deleting Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 940.689905] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-18597840-a092-459b-89ad-988b1290f0c0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.699226] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 940.699226] env[63293]: value = "task-1327950" [ 940.699226] env[63293]: _type = "Task" [ 940.699226] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.714902] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327950, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.763560] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327947, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.848118] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2bc0a3f8-bd65-4908-bb85-ac1bf77001eb tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "fe55a9b7-28b5-4b04-84f1-6b85b579a144" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.028s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.906513] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.966588] env[63293]: DEBUG nova.compute.manager [req-8b61d464-bc62-4dd1-9b22-f9c3efc1f342 req-ba892964-64d5-47e3-b0da-fad207e55276 service nova] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Received event network-vif-deleted-e54443d3-84a6-4e40-8220-bb2079d7805c {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.991741] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327949, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092296} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.992400] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 940.993791] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e617b9bc-356c-4fb4-a269-32eafbc397be {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.028392] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 828180c0-b7f8-4666-9f5d-e6741ef6495a/828180c0-b7f8-4666-9f5d-e6741ef6495a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.028773] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b285ace3-413a-44c1-b5bb-6bf70335b0a6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.052985] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 941.052985] env[63293]: value = "task-1327951" [ 941.052985] env[63293]: _type = "Task" [ 941.052985] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.066025] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327951, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.084437] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "f9ccb216-1435-44c5-ab34-a6388d794551" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.085123] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "f9ccb216-1435-44c5-ab34-a6388d794551" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.092609] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327946, 'name': ReconfigVM_Task, 'duration_secs': 1.04401} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.092609] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535/2f55ac2a-63d1-4713-ab0c-47fb6cf7d535.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.092609] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-973f6803-b0d5-44dd-92f4-a584f697bd68 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.101734] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 941.101734] env[63293]: value = "task-1327952" [ 941.101734] env[63293]: _type = "Task" [ 941.101734] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.114763] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327952, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.211448] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327950, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.264132] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327947, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.522919] env[63293]: DEBUG nova.compute.manager [req-d959cb23-58a1-4599-950d-06bb24cfed7a req-2dbe2c25-c025-4700-a0f1-6a5a00cf86c3 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Received event network-changed-df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.523111] env[63293]: DEBUG nova.compute.manager [req-d959cb23-58a1-4599-950d-06bb24cfed7a req-2dbe2c25-c025-4700-a0f1-6a5a00cf86c3 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing instance network info cache due to event network-changed-df1c83c6-23d7-439b-a496-234c48738d81. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 941.523341] env[63293]: DEBUG oslo_concurrency.lockutils [req-d959cb23-58a1-4599-950d-06bb24cfed7a req-2dbe2c25-c025-4700-a0f1-6a5a00cf86c3 service nova] Acquiring lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.523489] env[63293]: DEBUG oslo_concurrency.lockutils [req-d959cb23-58a1-4599-950d-06bb24cfed7a req-2dbe2c25-c025-4700-a0f1-6a5a00cf86c3 service nova] Acquired lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.523654] env[63293]: DEBUG nova.network.neutron [req-d959cb23-58a1-4599-950d-06bb24cfed7a req-2dbe2c25-c025-4700-a0f1-6a5a00cf86c3 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing network info cache for port df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 941.561515] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e422c2c3-f6bf-4536-bfb2-14d743940e98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.574933] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c901cc-f648-47c3-88db-688f33bfd7f8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.582710] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327951, 'name': ReconfigVM_Task, 'duration_secs': 0.432163} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.582710] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 828180c0-b7f8-4666-9f5d-e6741ef6495a/828180c0-b7f8-4666-9f5d-e6741ef6495a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.583541] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-74b0c180-bb76-4e07-aa2a-5d26cb5d016e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.614285] env[63293]: DEBUG nova.compute.manager [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 941.627437] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51dc2fc-6483-48de-84b2-fe3a2a69f6c5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.628556] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 941.628556] env[63293]: value = "task-1327953" [ 941.628556] env[63293]: _type = "Task" [ 941.628556] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.638032] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327952, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.641056] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8210d6c6-8936-4b43-8cef-8b062bbf0853 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.647327] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327953, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.660791] env[63293]: DEBUG nova.compute.provider_tree [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.713473] env[63293]: DEBUG oslo_vmware.api [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327950, 'name': RemoveSnapshot_Task, 'duration_secs': 0.649245} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.713693] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Deleted Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 941.713895] env[63293]: INFO nova.compute.manager [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Took 17.35 seconds to snapshot the instance on the hypervisor. [ 941.765861] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327947, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.138214] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327952, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.142449] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.145825] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327953, 'name': Rename_Task, 'duration_secs': 0.173184} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.146297] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.146550] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5356c486-81e5-4dea-ab68-6a29789951b7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.154799] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 942.154799] env[63293]: value = "task-1327954" [ 942.154799] env[63293]: _type = "Task" [ 942.154799] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.165268] env[63293]: DEBUG nova.scheduler.client.report [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.177605] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.271256] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327947, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.281641] env[63293]: DEBUG nova.compute.manager [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Found 3 images (rotation: 2) {{(pid=63293) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 942.281825] env[63293]: DEBUG nova.compute.manager [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Rotating out 1 backups {{(pid=63293) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 942.281993] env[63293]: DEBUG nova.compute.manager [None req-4b71a660-d2f7-49af-91c0-4edfb1a7079d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Deleting image 43bec244-5978-48cb-96cf-849335c88aa8 {{(pid=63293) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 942.314060] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "66be1abd-c541-4b8a-8d72-98bde03b1888" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.314944] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "66be1abd-c541-4b8a-8d72-98bde03b1888" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.383524] env[63293]: DEBUG nova.network.neutron [req-d959cb23-58a1-4599-950d-06bb24cfed7a req-2dbe2c25-c025-4700-a0f1-6a5a00cf86c3 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updated VIF entry in instance network info cache for port df1c83c6-23d7-439b-a496-234c48738d81. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.384551] env[63293]: DEBUG nova.network.neutron [req-d959cb23-58a1-4599-950d-06bb24cfed7a req-2dbe2c25-c025-4700-a0f1-6a5a00cf86c3 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updating instance_info_cache with network_info: [{"id": "df1c83c6-23d7-439b-a496-234c48738d81", "address": "fa:16:3e:0f:9f:14", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1c83c6-23", "ovs_interfaceid": "df1c83c6-23d7-439b-a496-234c48738d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.633390] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327952, 'name': Rename_Task, 'duration_secs': 1.082639} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.633827] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.634192] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c70c0a5a-7be9-4c99-8c8e-90dccddf144a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.644230] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 942.644230] env[63293]: value = "task-1327955" [ 942.644230] env[63293]: _type = "Task" [ 942.644230] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.657760] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327955, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.667803] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327954, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.679351] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.679928] env[63293]: DEBUG nova.compute.manager [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 942.682781] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.701s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.686242] env[63293]: INFO nova.compute.claims [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.766030] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327947, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.905164} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.766387] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/dcac372d-af95-4395-b80b-840bcd2b2ed4/dcac372d-af95-4395-b80b-840bcd2b2ed4.vmdk to [datastore1] ad585ebb-2072-45df-b645-94c9fa93576b/ad585ebb-2072-45df-b645-94c9fa93576b.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 942.767298] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694d320c-bd30-483a-99d6-a8a36bcbaf05 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.796653] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] ad585ebb-2072-45df-b645-94c9fa93576b/ad585ebb-2072-45df-b645-94c9fa93576b.vmdk or device None with type streamOptimized {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.797427] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4027f328-e52f-459e-a386-089edb608a19 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.817659] env[63293]: DEBUG nova.compute.manager [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 942.822096] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 942.822096] env[63293]: value = "task-1327956" [ 942.822096] env[63293]: _type = "Task" [ 942.822096] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.831518] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327956, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.886501] env[63293]: DEBUG oslo_concurrency.lockutils [req-d959cb23-58a1-4599-950d-06bb24cfed7a req-2dbe2c25-c025-4700-a0f1-6a5a00cf86c3 service nova] Releasing lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.157012] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327955, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.167305] env[63293]: DEBUG oslo_vmware.api [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327954, 'name': PowerOnVM_Task, 'duration_secs': 0.516458} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.167737] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 943.168183] env[63293]: INFO nova.compute.manager [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Took 9.00 seconds to spawn the instance on the hypervisor. [ 943.168242] env[63293]: DEBUG nova.compute.manager [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.169193] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9465327f-5364-4b58-834f-21913abb32a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.189433] env[63293]: DEBUG nova.compute.utils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 943.192795] env[63293]: DEBUG nova.compute.manager [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 943.193060] env[63293]: DEBUG nova.network.neutron [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 943.243554] env[63293]: DEBUG nova.policy [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd40ed85d59b1470481ca4dff87e4832c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eec85c5992d64d62a90e290a1aa2d441', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 943.337701] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327956, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.344860] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.589262] env[63293]: DEBUG nova.network.neutron [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Successfully created port: 1649020e-b967-4b7a-8ac3-e7f9c312702d {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.656242] env[63293]: DEBUG oslo_vmware.api [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1327955, 'name': PowerOnVM_Task, 'duration_secs': 0.537499} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.656721] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 943.656837] env[63293]: INFO nova.compute.manager [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Took 11.88 seconds to spawn the instance on the hypervisor. [ 943.657064] env[63293]: DEBUG nova.compute.manager [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.657922] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7131e7d5-0ad4-416b-815d-c1af140b43dd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.693623] env[63293]: DEBUG nova.compute.manager [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 943.696461] env[63293]: INFO nova.compute.manager [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Took 18.38 seconds to build instance. [ 943.840194] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327956, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.942887] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e72125d-0d8e-45d2-b828-6ae2faa58d57 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.952990] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5e76ff-c9a0-4d5a-aeb2-b70214164817 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.986990] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-641225fc-6b33-4e32-980a-bf6e5d441190 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.996773] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9834412c-dd35-41fd-91a7-8e269f02c3bb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.013489] env[63293]: DEBUG nova.compute.provider_tree [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.180127] env[63293]: INFO nova.compute.manager [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Took 20.53 seconds to build instance. [ 944.205915] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4289e05d-d95f-40f5-9190-facf3b89afdd tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "828180c0-b7f8-4666-9f5d-e6741ef6495a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.896s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.336880] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327956, 'name': ReconfigVM_Task, 'duration_secs': 1.052154} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.337314] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Reconfigured VM instance instance-0000003f to attach disk [datastore1] ad585ebb-2072-45df-b645-94c9fa93576b/ad585ebb-2072-45df-b645-94c9fa93576b.vmdk or device None with type streamOptimized {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.338798] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_name': '/dev/sda', 'boot_index': 0, 'encryption_format': None, 'disk_bus': None, 'size': 0, 'encryption_secret_uuid': None, 'device_type': 'disk', 'encryption_options': None, 'encrypted': False, 'guest_format': None, 'image_id': '1427aa66-93c7-49c6-9e28-dc7fe851dced'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283793', 'volume_id': '658fbd49-f54e-404a-99c2-e7147e4a3700', 'name': 'volume-658fbd49-f54e-404a-99c2-e7147e4a3700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'ad585ebb-2072-45df-b645-94c9fa93576b', 'attached_at': '', 'detached_at': '', 'volume_id': '658fbd49-f54e-404a-99c2-e7147e4a3700', 'serial': '658fbd49-f54e-404a-99c2-e7147e4a3700'}, 'attachment_id': '36203fea-1792-4438-a344-28a30fa862bb', 'device_type': None, 'delete_on_termination': False, 'mount_device': '/dev/sdb', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=63293) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 944.339015] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Volume attach. Driver type: vmdk {{(pid=63293) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 944.339218] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283793', 'volume_id': '658fbd49-f54e-404a-99c2-e7147e4a3700', 'name': 'volume-658fbd49-f54e-404a-99c2-e7147e4a3700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'ad585ebb-2072-45df-b645-94c9fa93576b', 'attached_at': '', 'detached_at': '', 'volume_id': '658fbd49-f54e-404a-99c2-e7147e4a3700', 'serial': '658fbd49-f54e-404a-99c2-e7147e4a3700'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 944.340023] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d544180-0f9d-41aa-bcec-e551e0c30664 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.360532] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a2aaf0-5348-4e98-9569-c0b7b26ae83a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.386182] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] volume-658fbd49-f54e-404a-99c2-e7147e4a3700/volume-658fbd49-f54e-404a-99c2-e7147e4a3700.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.386572] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82352df4-128a-4c4d-a647-caf463e2981c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.407479] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 944.407479] env[63293]: value = "task-1327957" [ 944.407479] env[63293]: _type = "Task" [ 944.407479] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.418095] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327957, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.503143] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "657d5f21-b985-4e3c-8799-b3aae2b1c509" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.503143] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "657d5f21-b985-4e3c-8799-b3aae2b1c509" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.517571] env[63293]: DEBUG nova.scheduler.client.report [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.682911] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7705069c-7ab2-4a47-9655-6a5ada3cd3ac tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.047s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.708698] env[63293]: DEBUG nova.compute.manager [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 944.735955] env[63293]: DEBUG nova.virt.hardware [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.736261] env[63293]: DEBUG nova.virt.hardware [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.736428] env[63293]: DEBUG nova.virt.hardware [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.736626] env[63293]: DEBUG nova.virt.hardware [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.736781] env[63293]: DEBUG nova.virt.hardware [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.736949] env[63293]: DEBUG nova.virt.hardware [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.737351] env[63293]: DEBUG nova.virt.hardware [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.737543] env[63293]: DEBUG nova.virt.hardware [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.737725] env[63293]: DEBUG nova.virt.hardware [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.737911] env[63293]: DEBUG nova.virt.hardware [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.738118] env[63293]: DEBUG nova.virt.hardware [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.739038] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047080d2-87cc-4c45-818d-e84cc4518af9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.748263] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff071f8-4c35-49df-8634-73e7ed391a16 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.918736] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327957, 'name': ReconfigVM_Task, 'duration_secs': 0.509406} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.920033] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Reconfigured VM instance instance-0000003f to attach disk [datastore1] volume-658fbd49-f54e-404a-99c2-e7147e4a3700/volume-658fbd49-f54e-404a-99c2-e7147e4a3700.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.925554] env[63293]: DEBUG nova.compute.manager [req-5d887536-0f3e-4b20-af94-e1220d036171 req-64960360-d101-4a5e-9514-a1fa2a2d29fe service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Received event network-changed-9e8240fa-3546-43eb-9cdc-4b42859690b3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.925749] env[63293]: DEBUG nova.compute.manager [req-5d887536-0f3e-4b20-af94-e1220d036171 req-64960360-d101-4a5e-9514-a1fa2a2d29fe service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Refreshing instance network info cache due to event network-changed-9e8240fa-3546-43eb-9cdc-4b42859690b3. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 944.925966] env[63293]: DEBUG oslo_concurrency.lockutils [req-5d887536-0f3e-4b20-af94-e1220d036171 req-64960360-d101-4a5e-9514-a1fa2a2d29fe service nova] Acquiring lock "refresh_cache-2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.926127] env[63293]: DEBUG oslo_concurrency.lockutils [req-5d887536-0f3e-4b20-af94-e1220d036171 req-64960360-d101-4a5e-9514-a1fa2a2d29fe service nova] Acquired lock "refresh_cache-2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.926422] env[63293]: DEBUG nova.network.neutron [req-5d887536-0f3e-4b20-af94-e1220d036171 req-64960360-d101-4a5e-9514-a1fa2a2d29fe service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Refreshing network info cache for port 9e8240fa-3546-43eb-9cdc-4b42859690b3 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 944.927729] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65ccd901-c3ed-4a8e-ac96-fc74f0971e3c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.950707] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 944.950707] env[63293]: value = "task-1327958" [ 944.950707] env[63293]: _type = "Task" [ 944.950707] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.960039] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327958, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.005064] env[63293]: DEBUG nova.compute.manager [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 945.021484] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.021484] env[63293]: DEBUG nova.compute.manager [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 945.024652] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.621s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.025043] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.027571] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.398s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.027872] env[63293]: DEBUG nova.objects.instance [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lazy-loading 'resources' on Instance uuid eec9bb88-c5c4-4d8a-8a30-84b8d89e153d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.068609] env[63293]: INFO nova.scheduler.client.report [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted allocations for instance 76ebe2ec-332c-4d3a-957e-d0c928dafdc7 [ 945.215501] env[63293]: DEBUG nova.network.neutron [req-5d887536-0f3e-4b20-af94-e1220d036171 req-64960360-d101-4a5e-9514-a1fa2a2d29fe service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Updated VIF entry in instance network info cache for port 9e8240fa-3546-43eb-9cdc-4b42859690b3. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 945.216077] env[63293]: DEBUG nova.network.neutron [req-5d887536-0f3e-4b20-af94-e1220d036171 req-64960360-d101-4a5e-9514-a1fa2a2d29fe service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Updating instance_info_cache with network_info: [{"id": "9e8240fa-3546-43eb-9cdc-4b42859690b3", "address": "fa:16:3e:04:3a:e7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e8240fa-35", "ovs_interfaceid": "9e8240fa-3546-43eb-9cdc-4b42859690b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.289432] env[63293]: DEBUG nova.network.neutron [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Successfully updated port: 1649020e-b967-4b7a-8ac3-e7f9c312702d {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 945.363682] env[63293]: DEBUG oslo_concurrency.lockutils [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "9ddf4f23-1279-4cbd-8212-10f344060445" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.364059] env[63293]: DEBUG oslo_concurrency.lockutils [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.364267] env[63293]: DEBUG nova.compute.manager [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.365545] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-616b7c89-90ec-45fa-90c1-d825da8634d9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.373409] env[63293]: DEBUG nova.compute.manager [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63293) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 945.374058] env[63293]: DEBUG nova.objects.instance [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'flavor' on Instance uuid 9ddf4f23-1279-4cbd-8212-10f344060445 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.463068] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327958, 'name': ReconfigVM_Task, 'duration_secs': 0.186018} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.463068] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283793', 'volume_id': '658fbd49-f54e-404a-99c2-e7147e4a3700', 'name': 'volume-658fbd49-f54e-404a-99c2-e7147e4a3700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'ad585ebb-2072-45df-b645-94c9fa93576b', 'attached_at': '', 'detached_at': '', 'volume_id': '658fbd49-f54e-404a-99c2-e7147e4a3700', 'serial': '658fbd49-f54e-404a-99c2-e7147e4a3700'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 945.463492] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c19dcb17-2c5a-44be-9779-7ffe0a8c925d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.470459] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 945.470459] env[63293]: value = "task-1327959" [ 945.470459] env[63293]: _type = "Task" [ 945.470459] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.479145] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327959, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.525200] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.535135] env[63293]: DEBUG nova.compute.utils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 945.536952] env[63293]: DEBUG nova.compute.manager [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 945.536952] env[63293]: DEBUG nova.network.neutron [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 945.577100] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1da68046-f527-4ab8-a8d6-fb7adb3eced5 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "76ebe2ec-332c-4d3a-957e-d0c928dafdc7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.263s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.582268] env[63293]: DEBUG nova.policy [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6860a9e34a1b44029cf57d9f04ae87c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fac34e49d5dc49e7a7055c998d5b6766', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 945.718554] env[63293]: DEBUG oslo_concurrency.lockutils [req-5d887536-0f3e-4b20-af94-e1220d036171 req-64960360-d101-4a5e-9514-a1fa2a2d29fe service nova] Releasing lock "refresh_cache-2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.783541] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637f54a0-c286-4d30-9b95-061f518afe4e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.792453] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92cbf610-81f0-4b26-98ba-3f7d4befb832 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.796533] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "refresh_cache-6f1fb925-72c2-4309-b3fd-aea217d38bf2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.796682] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired lock "refresh_cache-6f1fb925-72c2-4309-b3fd-aea217d38bf2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.796847] env[63293]: DEBUG nova.network.neutron [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 945.827860] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee96106c-3586-4a87-bf70-62ecdbbe0266 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.837556] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4551d7-0ef1-4042-bbfd-e5e86c143869 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.857457] env[63293]: DEBUG nova.compute.provider_tree [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.880078] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.884435] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85989766-e336-48ce-b285-95fa93fd7289 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.889354] env[63293]: DEBUG oslo_vmware.api [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 945.889354] env[63293]: value = "task-1327960" [ 945.889354] env[63293]: _type = "Task" [ 945.889354] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.895555] env[63293]: DEBUG nova.network.neutron [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Successfully created port: a20688c6-3fbe-4355-85f9-334019e37f43 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 945.901371] env[63293]: DEBUG oslo_vmware.api [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.981550] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327959, 'name': Rename_Task, 'duration_secs': 0.204902} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.982327] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.982747] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73e48700-55c0-4825-b93a-73b8b620995b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.991159] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 945.991159] env[63293]: value = "task-1327961" [ 945.991159] env[63293]: _type = "Task" [ 945.991159] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.001809] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327961, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.042979] env[63293]: DEBUG nova.compute.manager [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 946.329072] env[63293]: DEBUG nova.network.neutron [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 946.360816] env[63293]: DEBUG nova.scheduler.client.report [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.402394] env[63293]: DEBUG oslo_vmware.api [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327960, 'name': PowerOffVM_Task, 'duration_secs': 0.22493} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.402684] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 946.402864] env[63293]: DEBUG nova.compute.manager [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.403682] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcdce153-55c1-49a9-be7a-4c41c17aab26 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.470020] env[63293]: DEBUG nova.network.neutron [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Updating instance_info_cache with network_info: [{"id": "1649020e-b967-4b7a-8ac3-e7f9c312702d", "address": "fa:16:3e:1d:b4:5f", "network": {"id": "8af42ccf-7ab8-424b-a290-8383c292ef85", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1850566934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eec85c5992d64d62a90e290a1aa2d441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1649020e-b9", "ovs_interfaceid": "1649020e-b967-4b7a-8ac3-e7f9c312702d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.502240] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327961, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.869088] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.841s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.871676] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.966s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.872179] env[63293]: DEBUG nova.objects.instance [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lazy-loading 'resources' on Instance uuid 2557bded-6ccb-4c0f-8b97-34529b4bb740 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.891360] env[63293]: INFO nova.scheduler.client.report [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Deleted allocations for instance eec9bb88-c5c4-4d8a-8a30-84b8d89e153d [ 946.915497] env[63293]: DEBUG oslo_concurrency.lockutils [None req-12b79715-c7a0-4144-8b3d-33a5521b65fe tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.551s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.957092] env[63293]: DEBUG nova.compute.manager [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Received event network-vif-plugged-1649020e-b967-4b7a-8ac3-e7f9c312702d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.957320] env[63293]: DEBUG oslo_concurrency.lockutils [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] Acquiring lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.957535] env[63293]: DEBUG oslo_concurrency.lockutils [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] Lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.957729] env[63293]: DEBUG oslo_concurrency.lockutils [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] Lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.958110] env[63293]: DEBUG nova.compute.manager [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] No waiting events found dispatching network-vif-plugged-1649020e-b967-4b7a-8ac3-e7f9c312702d {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 946.958316] env[63293]: WARNING nova.compute.manager [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Received unexpected event network-vif-plugged-1649020e-b967-4b7a-8ac3-e7f9c312702d for instance with vm_state building and task_state spawning. [ 946.958476] env[63293]: DEBUG nova.compute.manager [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Received event network-changed-1649020e-b967-4b7a-8ac3-e7f9c312702d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.958639] env[63293]: DEBUG nova.compute.manager [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Refreshing instance network info cache due to event network-changed-1649020e-b967-4b7a-8ac3-e7f9c312702d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.958820] env[63293]: DEBUG oslo_concurrency.lockutils [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] Acquiring lock "refresh_cache-6f1fb925-72c2-4309-b3fd-aea217d38bf2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.972921] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lock "refresh_cache-6f1fb925-72c2-4309-b3fd-aea217d38bf2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.973245] env[63293]: DEBUG nova.compute.manager [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Instance network_info: |[{"id": "1649020e-b967-4b7a-8ac3-e7f9c312702d", "address": "fa:16:3e:1d:b4:5f", "network": {"id": "8af42ccf-7ab8-424b-a290-8383c292ef85", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1850566934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eec85c5992d64d62a90e290a1aa2d441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1649020e-b9", "ovs_interfaceid": "1649020e-b967-4b7a-8ac3-e7f9c312702d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 946.973534] env[63293]: DEBUG oslo_concurrency.lockutils [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] Acquired lock "refresh_cache-6f1fb925-72c2-4309-b3fd-aea217d38bf2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.973719] env[63293]: DEBUG nova.network.neutron [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Refreshing network info cache for port 1649020e-b967-4b7a-8ac3-e7f9c312702d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.974918] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:b4:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd48f0ef6-34e5-44d4-8baf-4470ed96ce73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1649020e-b967-4b7a-8ac3-e7f9c312702d', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.983092] env[63293]: DEBUG oslo.service.loopingcall [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.986310] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.986786] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-940d5f37-9b7d-4989-b37d-a64d77158643 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.014187] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327961, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.015731] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 947.015731] env[63293]: value = "task-1327962" [ 947.015731] env[63293]: _type = "Task" [ 947.015731] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.024458] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327962, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.050773] env[63293]: DEBUG nova.compute.manager [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 947.077349] env[63293]: DEBUG nova.virt.hardware [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.077780] env[63293]: DEBUG nova.virt.hardware [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.078071] env[63293]: DEBUG nova.virt.hardware [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.078385] env[63293]: DEBUG nova.virt.hardware [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.078619] env[63293]: DEBUG nova.virt.hardware [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.078829] env[63293]: DEBUG nova.virt.hardware [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.079119] env[63293]: DEBUG nova.virt.hardware [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.079312] env[63293]: DEBUG nova.virt.hardware [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.079500] env[63293]: DEBUG nova.virt.hardware [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.079673] env[63293]: DEBUG nova.virt.hardware [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.079881] env[63293]: DEBUG nova.virt.hardware [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.080776] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d591031b-b078-4dc2-b045-56040a311207 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.092287] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e88fd2-3edb-4944-9c3c-e1e4cc2e3c09 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.244834] env[63293]: DEBUG nova.network.neutron [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Updated VIF entry in instance network info cache for port 1649020e-b967-4b7a-8ac3-e7f9c312702d. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 947.245236] env[63293]: DEBUG nova.network.neutron [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Updating instance_info_cache with network_info: [{"id": "1649020e-b967-4b7a-8ac3-e7f9c312702d", "address": "fa:16:3e:1d:b4:5f", "network": {"id": "8af42ccf-7ab8-424b-a290-8383c292ef85", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1850566934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eec85c5992d64d62a90e290a1aa2d441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1649020e-b9", "ovs_interfaceid": "1649020e-b967-4b7a-8ac3-e7f9c312702d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.400404] env[63293]: DEBUG oslo_concurrency.lockutils [None req-adaf3f4b-dd51-4d5a-a321-7fcd72021920 tempest-ServersAdminNegativeTestJSON-702450948 tempest-ServersAdminNegativeTestJSON-702450948-project-member] Lock "eec9bb88-c5c4-4d8a-8a30-84b8d89e153d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.775s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.516151] env[63293]: DEBUG oslo_vmware.api [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1327961, 'name': PowerOnVM_Task, 'duration_secs': 1.189972} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.519178] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.532968] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327962, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.538339] env[63293]: DEBUG nova.network.neutron [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Successfully updated port: a20688c6-3fbe-4355-85f9-334019e37f43 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.615796] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b6961c-a037-47e1-ae07-f279a35d5f19 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.624637] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb940d9-9515-4aa6-9e23-bee094092ec7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.658939] env[63293]: DEBUG nova.compute.manager [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.659909] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9335042a-4446-44c2-9f13-1dd1f2ef56c5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.665678] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50cfc9fc-883b-4b99-8afc-9ed04de43116 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.675965] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f0d34c-b92a-4b9d-a98a-264c7c5935bc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.694194] env[63293]: DEBUG nova.compute.provider_tree [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.747624] env[63293]: DEBUG oslo_concurrency.lockutils [req-222db6b7-c3ca-483b-bde9-7386eca4022f req-7c3313d8-558c-4b51-a990-9d6774c2c6ca service nova] Releasing lock "refresh_cache-6f1fb925-72c2-4309-b3fd-aea217d38bf2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.029802] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327962, 'name': CreateVM_Task, 'duration_secs': 0.61919} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.030136] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 948.030722] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.030966] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.031299] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 948.031563] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca59bdd0-81c7-43be-a5fe-df73c6420c32 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.037882] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 948.037882] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fb40a4-cd85-20e7-233b-e4c866893e08" [ 948.037882] env[63293]: _type = "Task" [ 948.037882] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.044552] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.044552] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.044552] env[63293]: DEBUG nova.network.neutron [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.049224] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fb40a4-cd85-20e7-233b-e4c866893e08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.386206] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a8014c18-8384-4f13-9551-b2f1b466b83e tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 36.773s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.386206] env[63293]: DEBUG nova.scheduler.client.report [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.552010] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fb40a4-cd85-20e7-233b-e4c866893e08, 'name': SearchDatastore_Task, 'duration_secs': 0.026363} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.552680] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.552965] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.553271] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.553432] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.553632] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.553895] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f6ed42c-14b1-4822-b33c-c60c25e0607f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.563348] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.563531] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 948.564370] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ef1cd18-4db7-4967-a64a-545c3ff649e1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.570159] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 948.570159] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520a7c10-b757-11d3-5e44-11d1a19caa37" [ 948.570159] env[63293]: _type = "Task" [ 948.570159] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.578229] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520a7c10-b757-11d3-5e44-11d1a19caa37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.580061] env[63293]: DEBUG nova.network.neutron [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 948.703333] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.706012] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.564s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.707800] env[63293]: INFO nova.compute.claims [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 948.735589] env[63293]: INFO nova.scheduler.client.report [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleted allocations for instance 2557bded-6ccb-4c0f-8b97-34529b4bb740 [ 948.740280] env[63293]: DEBUG nova.compute.manager [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Stashing vm_state: stopped {{(pid=63293) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 948.759929] env[63293]: DEBUG nova.network.neutron [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance_info_cache with network_info: [{"id": "a20688c6-3fbe-4355-85f9-334019e37f43", "address": "fa:16:3e:08:30:f7", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa20688c6-3f", "ovs_interfaceid": "a20688c6-3fbe-4355-85f9-334019e37f43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.983397] env[63293]: DEBUG nova.compute.manager [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Received event network-vif-plugged-a20688c6-3fbe-4355-85f9-334019e37f43 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.983628] env[63293]: DEBUG oslo_concurrency.lockutils [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] Acquiring lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.983844] env[63293]: DEBUG oslo_concurrency.lockutils [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.984161] env[63293]: DEBUG oslo_concurrency.lockutils [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.984393] env[63293]: DEBUG nova.compute.manager [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] No waiting events found dispatching network-vif-plugged-a20688c6-3fbe-4355-85f9-334019e37f43 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 948.984573] env[63293]: WARNING nova.compute.manager [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Received unexpected event network-vif-plugged-a20688c6-3fbe-4355-85f9-334019e37f43 for instance with vm_state building and task_state spawning. [ 948.984743] env[63293]: DEBUG nova.compute.manager [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Received event network-changed-a20688c6-3fbe-4355-85f9-334019e37f43 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.984904] env[63293]: DEBUG nova.compute.manager [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Refreshing instance network info cache due to event network-changed-a20688c6-3fbe-4355-85f9-334019e37f43. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 948.985126] env[63293]: DEBUG oslo_concurrency.lockutils [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] Acquiring lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.081366] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]520a7c10-b757-11d3-5e44-11d1a19caa37, 'name': SearchDatastore_Task, 'duration_secs': 0.011386} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.082196] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e82b34a-380c-431c-88b8-103ce91979d2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.087964] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 949.087964] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b3a2b4-62fe-d449-8e55-6e901fc24992" [ 949.087964] env[63293]: _type = "Task" [ 949.087964] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.095699] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b3a2b4-62fe-d449-8e55-6e901fc24992, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.243321] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53149936-bc04-442c-862a-3d78a35307d0 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "2557bded-6ccb-4c0f-8b97-34529b4bb740" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.535s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.257378] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.261648] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.261940] env[63293]: DEBUG nova.compute.manager [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Instance network_info: |[{"id": "a20688c6-3fbe-4355-85f9-334019e37f43", "address": "fa:16:3e:08:30:f7", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa20688c6-3f", "ovs_interfaceid": "a20688c6-3fbe-4355-85f9-334019e37f43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 949.262272] env[63293]: DEBUG oslo_concurrency.lockutils [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] Acquired lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.262454] env[63293]: DEBUG nova.network.neutron [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Refreshing network info cache for port a20688c6-3fbe-4355-85f9-334019e37f43 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.263610] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:30:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aec0089a-ff85-4bef-bad8-c84de39af71a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a20688c6-3fbe-4355-85f9-334019e37f43', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.271691] env[63293]: DEBUG oslo.service.loopingcall [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.272474] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 949.272708] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dfe68f32-1853-41b2-9211-a804465d7a23 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.294843] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.294843] env[63293]: value = "task-1327963" [ 949.294843] env[63293]: _type = "Task" [ 949.294843] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.303381] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327963, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.598818] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52b3a2b4-62fe-d449-8e55-6e901fc24992, 'name': SearchDatastore_Task, 'duration_secs': 0.04892} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.598997] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.599259] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 6f1fb925-72c2-4309-b3fd-aea217d38bf2/6f1fb925-72c2-4309-b3fd-aea217d38bf2.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 949.599523] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a00b499-d928-46ae-858a-80665fcf5690 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.606477] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 949.606477] env[63293]: value = "task-1327964" [ 949.606477] env[63293]: _type = "Task" [ 949.606477] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.615296] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327964, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.807967] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327963, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.967478] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf9ce12-d8a6-499d-a6af-14a1cb9000e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.973502] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.973907] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.974188] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.974416] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.974659] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.980051] env[63293]: INFO nova.compute.manager [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Terminating instance [ 949.982461] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71919258-2dca-4ef8-89ef-57d06d6db040 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.986455] env[63293]: DEBUG nova.compute.manager [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.986665] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.987556] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48451ee-74e4-4456-bdcc-d03455ac83ec {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.026757] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49db1ee4-80ac-4df3-8405-96c2f20794dd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.029556] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 950.029865] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ab68fbd-9113-41b2-8910-811fe445f396 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.043514] env[63293]: DEBUG oslo_vmware.api [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 950.043514] env[63293]: value = "task-1327965" [ 950.043514] env[63293]: _type = "Task" [ 950.043514] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.045292] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c67b057-9909-44ca-b3f4-db4ddfb33372 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.061851] env[63293]: DEBUG oslo_vmware.api [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327965, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.070872] env[63293]: DEBUG nova.compute.provider_tree [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.083642] env[63293]: DEBUG nova.network.neutron [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updated VIF entry in instance network info cache for port a20688c6-3fbe-4355-85f9-334019e37f43. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.084037] env[63293]: DEBUG nova.network.neutron [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance_info_cache with network_info: [{"id": "a20688c6-3fbe-4355-85f9-334019e37f43", "address": "fa:16:3e:08:30:f7", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa20688c6-3f", "ovs_interfaceid": "a20688c6-3fbe-4355-85f9-334019e37f43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.119093] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327964, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.306032] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327963, 'name': CreateVM_Task, 'duration_secs': 0.628776} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.306032] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 950.306384] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.306510] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.306834] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 950.307116] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf2eee47-1ec5-42e0-acc1-c825f1120945 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.312075] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 950.312075] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a9d1c5-0eda-4cfa-feb6-3180925774bf" [ 950.312075] env[63293]: _type = "Task" [ 950.312075] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.321382] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a9d1c5-0eda-4cfa-feb6-3180925774bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.479692] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.480061] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.554571] env[63293]: DEBUG oslo_vmware.api [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327965, 'name': PowerOffVM_Task, 'duration_secs': 0.263005} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.554820] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 950.554998] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 950.555280] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61e0abfe-ed8e-4a0c-b250-0580d99496a1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.574725] env[63293]: DEBUG nova.scheduler.client.report [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.588027] env[63293]: DEBUG oslo_concurrency.lockutils [req-94c6455e-3b03-4831-a6e3-eafd2a3c40ba req-10fab1ed-f501-4751-990d-c9fa1dec655e service nova] Releasing lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.619251] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327964, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592987} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.619536] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 6f1fb925-72c2-4309-b3fd-aea217d38bf2/6f1fb925-72c2-4309-b3fd-aea217d38bf2.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 950.619754] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 950.620260] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f059469a-bd0e-4aac-9558-bb20d24ced26 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.628134] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 950.628134] env[63293]: value = "task-1327967" [ 950.628134] env[63293]: _type = "Task" [ 950.628134] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.639927] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327967, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.650660] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 950.650925] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 950.651154] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleting the datastore file [datastore2] 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.651448] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1e8ab9b-5644-4c0a-99e8-79366779eb9e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.665075] env[63293]: DEBUG oslo_vmware.api [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 950.665075] env[63293]: value = "task-1327968" [ 950.665075] env[63293]: _type = "Task" [ 950.665075] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.674644] env[63293]: DEBUG oslo_vmware.api [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327968, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.826165] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a9d1c5-0eda-4cfa-feb6-3180925774bf, 'name': SearchDatastore_Task, 'duration_secs': 0.013208} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.826596] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.826901] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.827239] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.827448] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.827696] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.828073] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f28898f-8c0d-4d1f-a0ae-6c2cbd2576f3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.844225] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.844425] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 950.845175] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ea5929c-cb86-4168-a8f4-5be1f2685b18 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.851194] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 950.851194] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52890231-ebc6-7e26-e7c4-99ed602165ed" [ 950.851194] env[63293]: _type = "Task" [ 950.851194] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.860566] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52890231-ebc6-7e26-e7c4-99ed602165ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.983471] env[63293]: DEBUG nova.compute.manager [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 951.080486] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.081054] env[63293]: DEBUG nova.compute.manager [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 951.083893] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.739s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.085461] env[63293]: INFO nova.compute.claims [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.144669] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327967, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.18573} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.145267] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 951.146179] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ebd22f-5766-4921-849d-45646fad5689 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.172298] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 6f1fb925-72c2-4309-b3fd-aea217d38bf2/6f1fb925-72c2-4309-b3fd-aea217d38bf2.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.172689] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f406ae8-c54b-4c42-823a-9aaf077e41b6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.197515] env[63293]: DEBUG oslo_vmware.api [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1327968, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.32445} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.198793] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.198985] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 951.199185] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 951.199358] env[63293]: INFO nova.compute.manager [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Took 1.21 seconds to destroy the instance on the hypervisor. [ 951.199600] env[63293]: DEBUG oslo.service.loopingcall [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.199877] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 951.199877] env[63293]: value = "task-1327969" [ 951.199877] env[63293]: _type = "Task" [ 951.199877] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.200128] env[63293]: DEBUG nova.compute.manager [-] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 951.200169] env[63293]: DEBUG nova.network.neutron [-] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 951.210242] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327969, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.362586] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52890231-ebc6-7e26-e7c4-99ed602165ed, 'name': SearchDatastore_Task, 'duration_secs': 0.015805} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.363407] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68c27d33-026b-4e55-b886-ad2d24c651e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.369632] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 951.369632] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5261e0a1-d4d9-0111-d7ce-b9660ee7b6ff" [ 951.369632] env[63293]: _type = "Task" [ 951.369632] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.378688] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5261e0a1-d4d9-0111-d7ce-b9660ee7b6ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.460647] env[63293]: DEBUG nova.compute.manager [req-27309d0c-64ed-4c85-904e-3acd64bc1b25 req-a67e18a5-f183-47a2-acee-ad0d9b47bdcb service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Received event network-vif-deleted-5533b032-6f43-4c5e-92e7-da17c807dbc1 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.460848] env[63293]: INFO nova.compute.manager [req-27309d0c-64ed-4c85-904e-3acd64bc1b25 req-a67e18a5-f183-47a2-acee-ad0d9b47bdcb service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Neutron deleted interface 5533b032-6f43-4c5e-92e7-da17c807dbc1; detaching it from the instance and deleting it from the info cache [ 951.461061] env[63293]: DEBUG nova.network.neutron [req-27309d0c-64ed-4c85-904e-3acd64bc1b25 req-a67e18a5-f183-47a2-acee-ad0d9b47bdcb service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.505126] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.590420] env[63293]: DEBUG nova.compute.utils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 951.593657] env[63293]: DEBUG nova.compute.manager [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 951.593829] env[63293]: DEBUG nova.network.neutron [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 951.631856] env[63293]: DEBUG nova.policy [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2312d48038d7474481a0709267645780', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '866b347100454019b07f63922b995bb6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.713749] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327969, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.881393] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5261e0a1-d4d9-0111-d7ce-b9660ee7b6ff, 'name': SearchDatastore_Task, 'duration_secs': 0.010706} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.881717] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.882020] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a/f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 951.882536] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9cd3d8e5-cfb3-4d97-b4e8-f5522388bfbe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.889437] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 951.889437] env[63293]: value = "task-1327970" [ 951.889437] env[63293]: _type = "Task" [ 951.889437] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.897846] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327970, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.919104] env[63293]: DEBUG nova.network.neutron [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Successfully created port: 5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 951.937146] env[63293]: DEBUG nova.network.neutron [-] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.965258] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ec1a997e-bc99-4e04-8d1a-b4f43de1ecde {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.976167] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37209e0-353a-40bb-90ff-a72c7d7d5d19 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.008897] env[63293]: DEBUG nova.compute.manager [req-27309d0c-64ed-4c85-904e-3acd64bc1b25 req-a67e18a5-f183-47a2-acee-ad0d9b47bdcb service nova] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Detach interface failed, port_id=5533b032-6f43-4c5e-92e7-da17c807dbc1, reason: Instance 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 952.094340] env[63293]: DEBUG nova.compute.manager [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 952.219240] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327969, 'name': ReconfigVM_Task, 'duration_secs': 0.681512} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.219636] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 6f1fb925-72c2-4309-b3fd-aea217d38bf2/6f1fb925-72c2-4309-b3fd-aea217d38bf2.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.220423] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78ee150f-8f75-4b3a-b4a9-59ec1a519944 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.229516] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 952.229516] env[63293]: value = "task-1327971" [ 952.229516] env[63293]: _type = "Task" [ 952.229516] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.245975] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327971, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.357494] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a25c9ee-3061-41d9-9ff1-ffd35962ba7d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.371110] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342dddae-4918-4e55-aa3e-e2d4b959320c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.411785] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2890056f-2f57-46d3-92cc-fb7727fadf45 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.420443] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327970, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524579} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.422650] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a/f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 952.422876] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.423186] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b9733a1-5d94-4a56-8a1b-df1eca125d49 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.425911] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a705e62-58d6-409a-a6dc-1b5bb7994a50 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.441418] env[63293]: INFO nova.compute.manager [-] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Took 1.24 seconds to deallocate network for instance. [ 952.441952] env[63293]: DEBUG nova.compute.provider_tree [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.446524] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 952.446524] env[63293]: value = "task-1327972" [ 952.446524] env[63293]: _type = "Task" [ 952.446524] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.462165] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327972, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.743015] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327971, 'name': Rename_Task, 'duration_secs': 0.372987} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.743374] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.743666] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edffb671-6a89-48ba-8b7b-a9b17cdd5f1d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.750341] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 952.750341] env[63293]: value = "task-1327973" [ 952.750341] env[63293]: _type = "Task" [ 952.750341] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.758667] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327973, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.945653] env[63293]: DEBUG nova.scheduler.client.report [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.955620] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.958842] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327972, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06551} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.959114] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.959959] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fe2dce-3d45-4999-a0cd-0920ec5d32ac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.985325] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a/f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.985962] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59224b75-d6e0-479c-934f-8710130ca442 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.008106] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 953.008106] env[63293]: value = "task-1327974" [ 953.008106] env[63293]: _type = "Task" [ 953.008106] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.016640] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327974, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.108110] env[63293]: DEBUG nova.compute.manager [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 953.132545] env[63293]: DEBUG nova.virt.hardware [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 953.132806] env[63293]: DEBUG nova.virt.hardware [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 953.132978] env[63293]: DEBUG nova.virt.hardware [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 953.133176] env[63293]: DEBUG nova.virt.hardware [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 953.133327] env[63293]: DEBUG nova.virt.hardware [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 953.133477] env[63293]: DEBUG nova.virt.hardware [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 953.133683] env[63293]: DEBUG nova.virt.hardware [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 953.133848] env[63293]: DEBUG nova.virt.hardware [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 953.134033] env[63293]: DEBUG nova.virt.hardware [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 953.134367] env[63293]: DEBUG nova.virt.hardware [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 953.134563] env[63293]: DEBUG nova.virt.hardware [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 953.135475] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22162273-146f-4f10-aecd-77ff41dff63d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.144228] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6db9842-ab6e-4263-85bc-f732fd6db5cb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.261423] env[63293]: DEBUG oslo_vmware.api [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327973, 'name': PowerOnVM_Task, 'duration_secs': 0.474779} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.261743] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.261979] env[63293]: INFO nova.compute.manager [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Took 8.55 seconds to spawn the instance on the hypervisor. [ 953.262189] env[63293]: DEBUG nova.compute.manager [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.262983] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa547547-0779-4079-9692-55507600f254 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.453794] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.454333] env[63293]: DEBUG nova.compute.manager [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 953.456953] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.932s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.461087] env[63293]: INFO nova.compute.claims [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 953.500371] env[63293]: DEBUG nova.compute.manager [req-fb9d672b-512a-46d5-8e2c-7fd8aad7e7dc req-49fc63dd-530a-4cdf-aa6a-73f7aa880bfc service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Received event network-vif-plugged-5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.500640] env[63293]: DEBUG oslo_concurrency.lockutils [req-fb9d672b-512a-46d5-8e2c-7fd8aad7e7dc req-49fc63dd-530a-4cdf-aa6a-73f7aa880bfc service nova] Acquiring lock "f9ccb216-1435-44c5-ab34-a6388d794551-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.500810] env[63293]: DEBUG oslo_concurrency.lockutils [req-fb9d672b-512a-46d5-8e2c-7fd8aad7e7dc req-49fc63dd-530a-4cdf-aa6a-73f7aa880bfc service nova] Lock "f9ccb216-1435-44c5-ab34-a6388d794551-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.500994] env[63293]: DEBUG oslo_concurrency.lockutils [req-fb9d672b-512a-46d5-8e2c-7fd8aad7e7dc req-49fc63dd-530a-4cdf-aa6a-73f7aa880bfc service nova] Lock "f9ccb216-1435-44c5-ab34-a6388d794551-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.501303] env[63293]: DEBUG nova.compute.manager [req-fb9d672b-512a-46d5-8e2c-7fd8aad7e7dc req-49fc63dd-530a-4cdf-aa6a-73f7aa880bfc service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] No waiting events found dispatching network-vif-plugged-5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 953.501548] env[63293]: WARNING nova.compute.manager [req-fb9d672b-512a-46d5-8e2c-7fd8aad7e7dc req-49fc63dd-530a-4cdf-aa6a-73f7aa880bfc service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Received unexpected event network-vif-plugged-5d29fa7c-6329-4700-b988-ac08592d4df7 for instance with vm_state building and task_state spawning. [ 953.519215] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327974, 'name': ReconfigVM_Task, 'duration_secs': 0.299991} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.519477] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Reconfigured VM instance instance-0000005a to attach disk [datastore2] f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a/f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.520112] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9de3b37c-0628-467f-8427-05f41855af4b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.527338] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 953.527338] env[63293]: value = "task-1327975" [ 953.527338] env[63293]: _type = "Task" [ 953.527338] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.536894] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327975, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.585195] env[63293]: DEBUG nova.network.neutron [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Successfully updated port: 5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 953.784078] env[63293]: INFO nova.compute.manager [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Took 16.81 seconds to build instance. [ 953.962918] env[63293]: DEBUG nova.compute.utils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.967346] env[63293]: DEBUG nova.compute.manager [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 953.967601] env[63293]: DEBUG nova.network.neutron [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 954.013747] env[63293]: DEBUG nova.policy [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd2a9495ab0b4768a4075a71f25aa9a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff8479fc251544e9b4618deccf992754', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 954.038874] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327975, 'name': Rename_Task, 'duration_secs': 0.152463} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.039129] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 954.039388] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b1e25a9-f920-4e4d-90b7-8132509e243d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.047969] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 954.047969] env[63293]: value = "task-1327976" [ 954.047969] env[63293]: _type = "Task" [ 954.047969] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.057126] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327976, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.087270] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.087474] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.087574] env[63293]: DEBUG nova.network.neutron [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.285694] env[63293]: DEBUG oslo_concurrency.lockutils [None req-91a68733-9f08-4cd1-972c-a2fb8877bcae tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.329s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.314929] env[63293]: DEBUG nova.network.neutron [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Successfully created port: 3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.468097] env[63293]: DEBUG nova.compute.manager [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 954.563743] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327976, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.628743] env[63293]: DEBUG nova.network.neutron [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 954.732422] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0806eaff-ddd0-40f8-b659-3ba01333e007 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.740271] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4345cf70-52b9-46d1-81be-059644463f25 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.776204] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba4dfa2-5638-4d19-858e-40c2931ee0be {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.785062] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d31062-ab92-4d47-9413-322dfdc42e94 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.801995] env[63293]: DEBUG nova.compute.provider_tree [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.840880] env[63293]: DEBUG nova.network.neutron [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updating instance_info_cache with network_info: [{"id": "5d29fa7c-6329-4700-b988-ac08592d4df7", "address": "fa:16:3e:56:b9:c4", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d29fa7c-63", "ovs_interfaceid": "5d29fa7c-6329-4700-b988-ac08592d4df7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.931479] env[63293]: DEBUG oslo_concurrency.lockutils [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.931783] env[63293]: DEBUG oslo_concurrency.lockutils [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.932096] env[63293]: DEBUG oslo_concurrency.lockutils [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.932261] env[63293]: DEBUG oslo_concurrency.lockutils [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.932444] env[63293]: DEBUG oslo_concurrency.lockutils [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.935045] env[63293]: INFO nova.compute.manager [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Terminating instance [ 954.938500] env[63293]: DEBUG nova.compute.manager [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.938703] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.939535] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07281a7-6f6d-478c-86f9-d17f0621bf1d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.948019] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.948268] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09fa8a53-46ad-4721-97e6-9e7eae2bfe9a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.955108] env[63293]: DEBUG oslo_vmware.api [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 954.955108] env[63293]: value = "task-1327977" [ 954.955108] env[63293]: _type = "Task" [ 954.955108] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.963449] env[63293]: DEBUG oslo_vmware.api [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327977, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.063059] env[63293]: DEBUG oslo_vmware.api [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1327976, 'name': PowerOnVM_Task, 'duration_secs': 0.653274} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.063295] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 955.063599] env[63293]: INFO nova.compute.manager [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Took 8.01 seconds to spawn the instance on the hypervisor. [ 955.063786] env[63293]: DEBUG nova.compute.manager [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.064641] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff094f8-5941-441e-80c3-1bcf7cecfe75 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.180628] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.180862] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.181978] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.181978] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.181978] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.183863] env[63293]: INFO nova.compute.manager [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Terminating instance [ 955.186139] env[63293]: DEBUG nova.compute.manager [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 955.186368] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.187377] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428de020-4a91-4f2c-8230-1c9a0c9a640e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.196724] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.197034] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-148d60c5-e4ba-475f-9139-e77a656a13c3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.204814] env[63293]: DEBUG oslo_vmware.api [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 955.204814] env[63293]: value = "task-1327978" [ 955.204814] env[63293]: _type = "Task" [ 955.204814] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.214126] env[63293]: DEBUG oslo_vmware.api [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.305853] env[63293]: DEBUG nova.scheduler.client.report [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.344031] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.346903] env[63293]: DEBUG nova.compute.manager [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Instance network_info: |[{"id": "5d29fa7c-6329-4700-b988-ac08592d4df7", "address": "fa:16:3e:56:b9:c4", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d29fa7c-63", "ovs_interfaceid": "5d29fa7c-6329-4700-b988-ac08592d4df7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 955.346903] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:b9:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '459b8c74-0aa6-42b6-996a-42b1c5d7e5c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5d29fa7c-6329-4700-b988-ac08592d4df7', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 955.354745] env[63293]: DEBUG oslo.service.loopingcall [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.355099] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 955.358055] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-886e9525-9ca7-4334-897b-7eb208395233 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.379625] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 955.379625] env[63293]: value = "task-1327979" [ 955.379625] env[63293]: _type = "Task" [ 955.379625] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.394739] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327979, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.466183] env[63293]: DEBUG oslo_vmware.api [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327977, 'name': PowerOffVM_Task, 'duration_secs': 0.394935} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.466505] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.466682] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.466947] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79f9370b-ab7e-494a-bcf4-ff3eea4fbac7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.480711] env[63293]: DEBUG nova.compute.manager [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 955.507417] env[63293]: DEBUG nova.virt.hardware [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.507674] env[63293]: DEBUG nova.virt.hardware [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.507847] env[63293]: DEBUG nova.virt.hardware [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.508043] env[63293]: DEBUG nova.virt.hardware [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.508204] env[63293]: DEBUG nova.virt.hardware [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.508443] env[63293]: DEBUG nova.virt.hardware [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.508713] env[63293]: DEBUG nova.virt.hardware [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.508952] env[63293]: DEBUG nova.virt.hardware [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.509176] env[63293]: DEBUG nova.virt.hardware [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.509351] env[63293]: DEBUG nova.virt.hardware [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.509529] env[63293]: DEBUG nova.virt.hardware [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.510506] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6cc889-5ca3-40d4-b0b5-0974ac9cf1b5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.523054] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf69675-c72e-4ff6-91c0-db6f9cce4b21 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.539913] env[63293]: DEBUG nova.compute.manager [req-b7ad1b39-8ab0-4048-949c-37dccf79156c req-a9322356-7974-48c6-9c01-0bdfb73ab8da service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Received event network-changed-5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.540351] env[63293]: DEBUG nova.compute.manager [req-b7ad1b39-8ab0-4048-949c-37dccf79156c req-a9322356-7974-48c6-9c01-0bdfb73ab8da service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Refreshing instance network info cache due to event network-changed-5d29fa7c-6329-4700-b988-ac08592d4df7. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 955.540721] env[63293]: DEBUG oslo_concurrency.lockutils [req-b7ad1b39-8ab0-4048-949c-37dccf79156c req-a9322356-7974-48c6-9c01-0bdfb73ab8da service nova] Acquiring lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.540997] env[63293]: DEBUG oslo_concurrency.lockutils [req-b7ad1b39-8ab0-4048-949c-37dccf79156c req-a9322356-7974-48c6-9c01-0bdfb73ab8da service nova] Acquired lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.541316] env[63293]: DEBUG nova.network.neutron [req-b7ad1b39-8ab0-4048-949c-37dccf79156c req-a9322356-7974-48c6-9c01-0bdfb73ab8da service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Refreshing network info cache for port 5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 955.584782] env[63293]: INFO nova.compute.manager [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Took 16.63 seconds to build instance. [ 955.622186] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.622467] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.622632] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Deleting the datastore file [datastore1] 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.622925] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c714c73-60e4-4277-ba2a-821eeb20c8ad {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.632351] env[63293]: DEBUG oslo_vmware.api [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 955.632351] env[63293]: value = "task-1327981" [ 955.632351] env[63293]: _type = "Task" [ 955.632351] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.643750] env[63293]: DEBUG oslo_vmware.api [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327981, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.716653] env[63293]: DEBUG oslo_vmware.api [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327978, 'name': PowerOffVM_Task, 'duration_secs': 0.207937} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.716653] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.716653] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.716653] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f446e7c6-5dea-4d09-82fa-a063e69f7e0a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.787881] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.788406] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.788735] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Deleting the datastore file [datastore2] 6f1fb925-72c2-4309-b3fd-aea217d38bf2 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.789175] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ecee2826-ac26-472d-afa0-94f780bdb8e7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.797913] env[63293]: DEBUG oslo_vmware.api [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 955.797913] env[63293]: value = "task-1327983" [ 955.797913] env[63293]: _type = "Task" [ 955.797913] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.806041] env[63293]: DEBUG oslo_vmware.api [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327983, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.812076] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.812566] env[63293]: DEBUG nova.compute.manager [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 955.815326] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.558s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.891051] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327979, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.013449] env[63293]: DEBUG nova.network.neutron [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Successfully updated port: 3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 956.087066] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fa828dff-586a-43fa-9da7-0c70c40b1016 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.138s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.142834] env[63293]: DEBUG oslo_vmware.api [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1327981, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268386} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.143189] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.143455] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.143674] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.143847] env[63293]: INFO nova.compute.manager [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Took 1.21 seconds to destroy the instance on the hypervisor. [ 956.144102] env[63293]: DEBUG oslo.service.loopingcall [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.146430] env[63293]: DEBUG nova.compute.manager [-] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.146543] env[63293]: DEBUG nova.network.neutron [-] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.242408] env[63293]: DEBUG nova.network.neutron [req-b7ad1b39-8ab0-4048-949c-37dccf79156c req-a9322356-7974-48c6-9c01-0bdfb73ab8da service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updated VIF entry in instance network info cache for port 5d29fa7c-6329-4700-b988-ac08592d4df7. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 956.242759] env[63293]: DEBUG nova.network.neutron [req-b7ad1b39-8ab0-4048-949c-37dccf79156c req-a9322356-7974-48c6-9c01-0bdfb73ab8da service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updating instance_info_cache with network_info: [{"id": "5d29fa7c-6329-4700-b988-ac08592d4df7", "address": "fa:16:3e:56:b9:c4", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d29fa7c-63", "ovs_interfaceid": "5d29fa7c-6329-4700-b988-ac08592d4df7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.308842] env[63293]: DEBUG oslo_vmware.api [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1327983, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.499293} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.308842] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.309497] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.309497] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.309497] env[63293]: INFO nova.compute.manager [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Took 1.12 seconds to destroy the instance on the hypervisor. [ 956.313025] env[63293]: DEBUG oslo.service.loopingcall [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.313025] env[63293]: DEBUG nova.compute.manager [-] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.313025] env[63293]: DEBUG nova.network.neutron [-] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.320263] env[63293]: INFO nova.compute.claims [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.324420] env[63293]: DEBUG nova.compute.utils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 956.326030] env[63293]: DEBUG nova.compute.manager [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 956.326158] env[63293]: DEBUG nova.network.neutron [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 956.389513] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327979, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.390961] env[63293]: DEBUG nova.policy [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b94ec4d2961a43ca9a331a6051fa774d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7744835de0bc4e30b1b29ee5a439cd0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 956.518305] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-66be1abd-c541-4b8a-8d72-98bde03b1888" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.518305] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-66be1abd-c541-4b8a-8d72-98bde03b1888" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.518305] env[63293]: DEBUG nova.network.neutron [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 956.746035] env[63293]: DEBUG oslo_concurrency.lockutils [req-b7ad1b39-8ab0-4048-949c-37dccf79156c req-a9322356-7974-48c6-9c01-0bdfb73ab8da service nova] Releasing lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.763790] env[63293]: DEBUG nova.compute.manager [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Stashing vm_state: active {{(pid=63293) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 956.828192] env[63293]: INFO nova.compute.resource_tracker [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating resource usage from migration a66e3649-f4ed-4fb1-8bca-bb060f430e05 [ 956.831060] env[63293]: DEBUG nova.compute.manager [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 956.844030] env[63293]: DEBUG nova.compute.manager [req-afb7b7cb-f7ce-4243-ba41-5e1d2db764c1 req-c1d5a373-60ac-4ac0-af68-9c03950d452c service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Received event network-vif-deleted-2f0bfb1c-6184-49f1-9c71-cb7459e1470d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.844173] env[63293]: INFO nova.compute.manager [req-afb7b7cb-f7ce-4243-ba41-5e1d2db764c1 req-c1d5a373-60ac-4ac0-af68-9c03950d452c service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Neutron deleted interface 2f0bfb1c-6184-49f1-9c71-cb7459e1470d; detaching it from the instance and deleting it from the info cache [ 956.847113] env[63293]: DEBUG nova.network.neutron [req-afb7b7cb-f7ce-4243-ba41-5e1d2db764c1 req-c1d5a373-60ac-4ac0-af68-9c03950d452c service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.859363] env[63293]: DEBUG nova.network.neutron [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Successfully created port: 0a09da68-14b4-4caa-9967-6b00717e08d7 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 956.892068] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327979, 'name': CreateVM_Task, 'duration_secs': 1.429933} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.894641] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 956.895616] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.895783] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.896108] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 956.896364] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b23a50a-e02e-41fb-b917-3eac64e8b5bc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.904821] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 956.904821] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52dce41f-d385-0faa-4809-4b458ada180f" [ 956.904821] env[63293]: _type = "Task" [ 956.904821] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.918437] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52dce41f-d385-0faa-4809-4b458ada180f, 'name': SearchDatastore_Task, 'duration_secs': 0.010106} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.918897] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.919168] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 956.919400] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.922460] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.922460] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 956.922460] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edabe5c5-5675-4524-bcd6-b4e2343df157 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.932461] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 956.932679] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 956.933430] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04da1c3d-dd24-4a81-8af5-ad1fbec4df36 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.942145] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 956.942145] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5290a781-e820-edf0-49f1-ea2929f969f3" [ 956.942145] env[63293]: _type = "Task" [ 956.942145] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.953695] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5290a781-e820-edf0-49f1-ea2929f969f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.043065] env[63293]: DEBUG nova.network.neutron [-] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.089252] env[63293]: DEBUG nova.network.neutron [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 957.114047] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9640852b-eea1-4213-8ea9-4765f5c28e35 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.125032] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80466fd-c9ae-4d39-81e3-3e47f7fdc725 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.162272] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c6a5b7-8384-446d-9c12-077188c76387 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.173067] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9d5770-0952-4a4c-a03b-9905ebc3bb60 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.190584] env[63293]: DEBUG nova.compute.provider_tree [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.267897] env[63293]: DEBUG nova.network.neutron [-] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.286883] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.352811] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4afae35-5405-4a72-a2f6-e205f79b0167 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.357017] env[63293]: DEBUG nova.network.neutron [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Updating instance_info_cache with network_info: [{"id": "3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf", "address": "fa:16:3e:a0:fd:20", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fccaa93-b9", "ovs_interfaceid": "3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.362747] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b058fd-2c78-497b-9a94-376998061705 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.398901] env[63293]: DEBUG nova.compute.manager [req-afb7b7cb-f7ce-4243-ba41-5e1d2db764c1 req-c1d5a373-60ac-4ac0-af68-9c03950d452c service nova] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Detach interface failed, port_id=2f0bfb1c-6184-49f1-9c71-cb7459e1470d, reason: Instance 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 957.453248] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5290a781-e820-edf0-49f1-ea2929f969f3, 'name': SearchDatastore_Task, 'duration_secs': 0.00896} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.454074] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38cd908e-2fbc-48f2-93c6-72ddb1c50318 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.459566] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 957.459566] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a6e76f-49d5-a490-bbf2-7b5466749aa4" [ 957.459566] env[63293]: _type = "Task" [ 957.459566] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.467896] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a6e76f-49d5-a490-bbf2-7b5466749aa4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.546676] env[63293]: INFO nova.compute.manager [-] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Took 1.24 seconds to deallocate network for instance. [ 957.566935] env[63293]: DEBUG nova.compute.manager [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Received event network-vif-plugged-3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.567307] env[63293]: DEBUG oslo_concurrency.lockutils [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] Acquiring lock "66be1abd-c541-4b8a-8d72-98bde03b1888-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.567390] env[63293]: DEBUG oslo_concurrency.lockutils [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] Lock "66be1abd-c541-4b8a-8d72-98bde03b1888-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.567637] env[63293]: DEBUG oslo_concurrency.lockutils [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] Lock "66be1abd-c541-4b8a-8d72-98bde03b1888-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.567708] env[63293]: DEBUG nova.compute.manager [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] No waiting events found dispatching network-vif-plugged-3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 957.568670] env[63293]: WARNING nova.compute.manager [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Received unexpected event network-vif-plugged-3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf for instance with vm_state building and task_state spawning. [ 957.568670] env[63293]: DEBUG nova.compute.manager [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Received event network-changed-3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.568670] env[63293]: DEBUG nova.compute.manager [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Refreshing instance network info cache due to event network-changed-3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 957.568670] env[63293]: DEBUG oslo_concurrency.lockutils [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] Acquiring lock "refresh_cache-66be1abd-c541-4b8a-8d72-98bde03b1888" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.696692] env[63293]: DEBUG nova.scheduler.client.report [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.773604] env[63293]: INFO nova.compute.manager [-] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Took 1.63 seconds to deallocate network for instance. [ 957.840963] env[63293]: DEBUG nova.compute.manager [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 957.860248] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-66be1abd-c541-4b8a-8d72-98bde03b1888" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.860682] env[63293]: DEBUG nova.compute.manager [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Instance network_info: |[{"id": "3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf", "address": "fa:16:3e:a0:fd:20", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fccaa93-b9", "ovs_interfaceid": "3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 957.860897] env[63293]: DEBUG oslo_concurrency.lockutils [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] Acquired lock "refresh_cache-66be1abd-c541-4b8a-8d72-98bde03b1888" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.861698] env[63293]: DEBUG nova.network.neutron [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Refreshing network info cache for port 3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 957.863039] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:fd:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.870640] env[63293]: DEBUG oslo.service.loopingcall [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.873258] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 957.873585] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f889a56f-7556-458d-b272-357cc54eb58e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.890615] env[63293]: DEBUG nova.virt.hardware [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 957.890851] env[63293]: DEBUG nova.virt.hardware [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 957.891104] env[63293]: DEBUG nova.virt.hardware [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 957.891324] env[63293]: DEBUG nova.virt.hardware [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 957.891482] env[63293]: DEBUG nova.virt.hardware [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 957.891635] env[63293]: DEBUG nova.virt.hardware [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.891847] env[63293]: DEBUG nova.virt.hardware [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.892165] env[63293]: DEBUG nova.virt.hardware [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.892209] env[63293]: DEBUG nova.virt.hardware [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.892417] env[63293]: DEBUG nova.virt.hardware [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.892642] env[63293]: DEBUG nova.virt.hardware [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.894045] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ccb0ee-cf72-4f3b-912e-e5ea57c83b88 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.904014] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61325237-814b-4138-9239-cf88faf5455f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.908689] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.908689] env[63293]: value = "task-1327984" [ 957.908689] env[63293]: _type = "Task" [ 957.908689] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.925644] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327984, 'name': CreateVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.970816] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a6e76f-49d5-a490-bbf2-7b5466749aa4, 'name': SearchDatastore_Task, 'duration_secs': 0.010265} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.971124] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.971510] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] f9ccb216-1435-44c5-ab34-a6388d794551/f9ccb216-1435-44c5-ab34-a6388d794551.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 957.971826] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb976683-029c-4999-9064-990202dcd59f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.979794] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 957.979794] env[63293]: value = "task-1327985" [ 957.979794] env[63293]: _type = "Task" [ 957.979794] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.987925] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327985, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.052964] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.202149] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.387s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.202381] env[63293]: INFO nova.compute.manager [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Migrating [ 958.209409] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.704s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.210988] env[63293]: INFO nova.compute.claims [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.280705] env[63293]: DEBUG oslo_concurrency.lockutils [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.420748] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327984, 'name': CreateVM_Task, 'duration_secs': 0.384599} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.421164] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.424321] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.424405] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.424694] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.424966] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49dd2c49-4705-4551-9561-badc94b45e41 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.462435] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 958.462435] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525a5970-3e2f-7233-cc3f-232bc3b81525" [ 958.462435] env[63293]: _type = "Task" [ 958.462435] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.462435] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525a5970-3e2f-7233-cc3f-232bc3b81525, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.492678] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327985, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.725395] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.725582] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.725813] env[63293]: DEBUG nova.network.neutron [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 958.728960] env[63293]: DEBUG nova.network.neutron [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Updated VIF entry in instance network info cache for port 3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 958.729317] env[63293]: DEBUG nova.network.neutron [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Updating instance_info_cache with network_info: [{"id": "3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf", "address": "fa:16:3e:a0:fd:20", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fccaa93-b9", "ovs_interfaceid": "3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.797868] env[63293]: DEBUG nova.network.neutron [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Successfully updated port: 0a09da68-14b4-4caa-9967-6b00717e08d7 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 958.876513] env[63293]: DEBUG nova.compute.manager [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Received event network-vif-plugged-0a09da68-14b4-4caa-9967-6b00717e08d7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.876513] env[63293]: DEBUG oslo_concurrency.lockutils [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] Acquiring lock "657d5f21-b985-4e3c-8799-b3aae2b1c509-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.876513] env[63293]: DEBUG oslo_concurrency.lockutils [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] Lock "657d5f21-b985-4e3c-8799-b3aae2b1c509-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.876661] env[63293]: DEBUG oslo_concurrency.lockutils [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] Lock "657d5f21-b985-4e3c-8799-b3aae2b1c509-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.876843] env[63293]: DEBUG nova.compute.manager [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] No waiting events found dispatching network-vif-plugged-0a09da68-14b4-4caa-9967-6b00717e08d7 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 958.877166] env[63293]: WARNING nova.compute.manager [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Received unexpected event network-vif-plugged-0a09da68-14b4-4caa-9967-6b00717e08d7 for instance with vm_state building and task_state spawning. [ 958.877393] env[63293]: DEBUG nova.compute.manager [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Received event network-changed-0a09da68-14b4-4caa-9967-6b00717e08d7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.877566] env[63293]: DEBUG nova.compute.manager [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Refreshing instance network info cache due to event network-changed-0a09da68-14b4-4caa-9967-6b00717e08d7. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 958.877777] env[63293]: DEBUG oslo_concurrency.lockutils [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] Acquiring lock "refresh_cache-657d5f21-b985-4e3c-8799-b3aae2b1c509" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.877980] env[63293]: DEBUG oslo_concurrency.lockutils [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] Acquired lock "refresh_cache-657d5f21-b985-4e3c-8799-b3aae2b1c509" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.878210] env[63293]: DEBUG nova.network.neutron [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Refreshing network info cache for port 0a09da68-14b4-4caa-9967-6b00717e08d7 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 958.947692] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525a5970-3e2f-7233-cc3f-232bc3b81525, 'name': SearchDatastore_Task, 'duration_secs': 0.055598} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.948033] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.948272] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.949054] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.949054] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.949054] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.949264] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f5ff894-222a-4db3-a3ea-3bf4d67a5d71 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.963798] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.963987] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 958.964775] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f3b1dcb-269b-4e1b-b061-08ea4927a35c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.970624] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 958.970624] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527f4e98-6b0a-b217-f8a4-4c04981b6c52" [ 958.970624] env[63293]: _type = "Task" [ 958.970624] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.979192] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527f4e98-6b0a-b217-f8a4-4c04981b6c52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.989738] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327985, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.63794} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.991040] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] f9ccb216-1435-44c5-ab34-a6388d794551/f9ccb216-1435-44c5-ab34-a6388d794551.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 958.991040] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 958.991040] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-193d1a28-322b-4237-815d-05e023ec4f4f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.998165] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 958.998165] env[63293]: value = "task-1327986" [ 958.998165] env[63293]: _type = "Task" [ 958.998165] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.006697] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327986, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.233663] env[63293]: DEBUG oslo_concurrency.lockutils [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] Releasing lock "refresh_cache-66be1abd-c541-4b8a-8d72-98bde03b1888" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.234033] env[63293]: DEBUG nova.compute.manager [req-5f8faa0d-5163-46a4-ad88-af9d0ba08f29 req-bb4b10ef-07f2-44ae-ad08-908dfd089a51 service nova] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Received event network-vif-deleted-1649020e-b967-4b7a-8ac3-e7f9c312702d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.302625] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "refresh_cache-657d5f21-b985-4e3c-8799-b3aae2b1c509" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.425932] env[63293]: DEBUG nova.network.neutron [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 959.484445] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527f4e98-6b0a-b217-f8a4-4c04981b6c52, 'name': SearchDatastore_Task, 'duration_secs': 0.039103} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.489626] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-413405f4-70d4-4aa7-b368-d2748639e81d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.492628] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a992ad-d2d8-45d8-97e6-79f1d1535ea5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.499103] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 959.499103] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fc097c-1b5f-7c5c-d3e7-a18aba3b0740" [ 959.499103] env[63293]: _type = "Task" [ 959.499103] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.507048] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72408338-51ac-4938-b21c-70f52a0d9bfd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.515898] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327986, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066449} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.544804] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 959.545388] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fc097c-1b5f-7c5c-d3e7-a18aba3b0740, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.546116] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d15261-c017-4204-b2f0-805afbcc5626 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.549061] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd6d0bd-1cf8-40da-89f5-d8c278e7c701 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.552150] env[63293]: DEBUG nova.network.neutron [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance_info_cache with network_info: [{"id": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "address": "fa:16:3e:93:d3:25", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dbccf6a-05", "ovs_interfaceid": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.556040] env[63293]: DEBUG nova.network.neutron [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.571573] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827acc43-a5cd-4a9b-8e52-fe482a9dd622 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.584485] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] f9ccb216-1435-44c5-ab34-a6388d794551/f9ccb216-1435-44c5-ab34-a6388d794551.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.585833] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e26927a-d9a2-4c4b-a56f-b8baf439c884 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.611657] env[63293]: DEBUG nova.compute.provider_tree [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.614857] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 959.614857] env[63293]: value = "task-1327987" [ 959.614857] env[63293]: _type = "Task" [ 959.614857] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.624417] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327987, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.020097] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fc097c-1b5f-7c5c-d3e7-a18aba3b0740, 'name': SearchDatastore_Task, 'duration_secs': 0.056018} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.020448] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.020710] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 66be1abd-c541-4b8a-8d72-98bde03b1888/66be1abd-c541-4b8a-8d72-98bde03b1888.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 960.020997] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce7d4adb-44ab-46bc-866c-786a2fe31709 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.032480] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 960.032480] env[63293]: value = "task-1327988" [ 960.032480] env[63293]: _type = "Task" [ 960.032480] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.042218] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327988, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.054902] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.056429] env[63293]: DEBUG oslo_concurrency.lockutils [req-fbcc2493-131c-49e6-8ea4-b243583041d2 req-0e9851ff-7d8b-4c21-8e1c-84db982fa640 service nova] Releasing lock "refresh_cache-657d5f21-b985-4e3c-8799-b3aae2b1c509" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.056950] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "refresh_cache-657d5f21-b985-4e3c-8799-b3aae2b1c509" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.057121] env[63293]: DEBUG nova.network.neutron [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 960.115945] env[63293]: DEBUG nova.scheduler.client.report [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.129240] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327987, 'name': ReconfigVM_Task, 'duration_secs': 0.27861} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.129565] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Reconfigured VM instance instance-0000005b to attach disk [datastore2] f9ccb216-1435-44c5-ab34-a6388d794551/f9ccb216-1435-44c5-ab34-a6388d794551.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 960.130282] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7bf7010b-cb0e-4f40-adcf-ac81d29495ed {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.138116] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 960.138116] env[63293]: value = "task-1327989" [ 960.138116] env[63293]: _type = "Task" [ 960.138116] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.150470] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327989, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.543257] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327988, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487217} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.543592] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 66be1abd-c541-4b8a-8d72-98bde03b1888/66be1abd-c541-4b8a-8d72-98bde03b1888.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.543758] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.544027] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8bb1389c-43be-4d60-a42a-011347280df8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.551468] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 960.551468] env[63293]: value = "task-1327990" [ 960.551468] env[63293]: _type = "Task" [ 960.551468] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.567144] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327990, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.594765] env[63293]: DEBUG nova.network.neutron [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 960.624050] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.624584] env[63293]: DEBUG nova.compute.manager [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 960.627976] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.672s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.628150] env[63293]: DEBUG nova.objects.instance [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lazy-loading 'resources' on Instance uuid 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.651418] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327989, 'name': Rename_Task, 'duration_secs': 0.239836} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.651696] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 960.651946] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2299cd07-3cb3-4b25-a8f2-8087836951ad {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.659265] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 960.659265] env[63293]: value = "task-1327991" [ 960.659265] env[63293]: _type = "Task" [ 960.659265] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.671075] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327991, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.747791] env[63293]: DEBUG nova.network.neutron [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Updating instance_info_cache with network_info: [{"id": "0a09da68-14b4-4caa-9967-6b00717e08d7", "address": "fa:16:3e:3b:a7:78", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a09da68-14", "ovs_interfaceid": "0a09da68-14b4-4caa-9967-6b00717e08d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.061326] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327990, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06759} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.061650] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 961.062469] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28783b3-007a-4c46-bc60-a61d33b73387 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.087054] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 66be1abd-c541-4b8a-8d72-98bde03b1888/66be1abd-c541-4b8a-8d72-98bde03b1888.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.088416] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a498a07d-d07e-4432-be1c-2d27275ccf25 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.110079] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 961.110079] env[63293]: value = "task-1327992" [ 961.110079] env[63293]: _type = "Task" [ 961.110079] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.119476] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327992, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.131570] env[63293]: DEBUG nova.compute.utils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 961.136530] env[63293]: DEBUG nova.compute.manager [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 961.136855] env[63293]: DEBUG nova.network.neutron [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 961.170177] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327991, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.208920] env[63293]: DEBUG nova.policy [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc557354560049119144c62dd61d0816', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2af826d819b4cc4ab27ff8b95144039', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 961.250663] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "refresh_cache-657d5f21-b985-4e3c-8799-b3aae2b1c509" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.251013] env[63293]: DEBUG nova.compute.manager [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Instance network_info: |[{"id": "0a09da68-14b4-4caa-9967-6b00717e08d7", "address": "fa:16:3e:3b:a7:78", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a09da68-14", "ovs_interfaceid": "0a09da68-14b4-4caa-9967-6b00717e08d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 961.251654] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:a7:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ef746c57-cd18-4883-a0e9-c52937aaf41d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0a09da68-14b4-4caa-9967-6b00717e08d7', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.260123] env[63293]: DEBUG oslo.service.loopingcall [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.260367] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 961.260598] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3150dbd6-1eb6-40f9-9250-1ae1f607a272 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.286985] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.286985] env[63293]: value = "task-1327993" [ 961.286985] env[63293]: _type = "Task" [ 961.286985] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.296749] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327993, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.396334] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d8f992-1680-4467-97e1-f09c149034d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.406657] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a936b42c-9857-4909-98bf-47bbce9f42cd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.441319] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5f198d-6480-4c15-84e1-d2f014e2ee00 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.451326] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78109a1c-2095-4673-99ea-30d32ed24656 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.466346] env[63293]: DEBUG nova.compute.provider_tree [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.545340] env[63293]: DEBUG nova.network.neutron [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Successfully created port: bda9acec-71cb-4de2-859e-75837898a336 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.593624] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb4fb4d-8943-4339-9de9-15b5aec3b093 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.618325] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance '9ddf4f23-1279-4cbd-8212-10f344060445' progress to 0 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 961.631139] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327992, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.636828] env[63293]: DEBUG nova.compute.manager [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 961.676517] env[63293]: DEBUG oslo_vmware.api [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1327991, 'name': PowerOnVM_Task, 'duration_secs': 0.599904} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.676845] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 961.677095] env[63293]: INFO nova.compute.manager [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Took 8.57 seconds to spawn the instance on the hypervisor. [ 961.677324] env[63293]: DEBUG nova.compute.manager [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 961.678703] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476b71c5-2948-4486-bab5-4f2cf3933fe3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.797912] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327993, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.972619] env[63293]: DEBUG nova.scheduler.client.report [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.129836] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 962.129836] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327992, 'name': ReconfigVM_Task, 'duration_secs': 0.620774} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.129836] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7942bb4a-4761-4ce9-b127-0b6a7ba5a136 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.131439] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 66be1abd-c541-4b8a-8d72-98bde03b1888/66be1abd-c541-4b8a-8d72-98bde03b1888.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.132046] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3740fd91-d0f2-4df3-8d33-f77c1dfc2c89 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.143218] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 962.143218] env[63293]: value = "task-1327995" [ 962.143218] env[63293]: _type = "Task" [ 962.143218] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.144813] env[63293]: DEBUG oslo_vmware.api [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 962.144813] env[63293]: value = "task-1327994" [ 962.144813] env[63293]: _type = "Task" [ 962.144813] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.157926] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327995, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.161735] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] VM already powered off {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 962.161985] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance '9ddf4f23-1279-4cbd-8212-10f344060445' progress to 17 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 962.199273] env[63293]: INFO nova.compute.manager [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Took 20.08 seconds to build instance. [ 962.298101] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1327993, 'name': CreateVM_Task, 'duration_secs': 0.534119} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.298365] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 962.299106] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.299330] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.299674] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.299952] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25650098-79c8-4f62-bbe5-cca26d0a936b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.305717] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 962.305717] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523f169a-1b94-082c-c8d3-340b88b46677" [ 962.305717] env[63293]: _type = "Task" [ 962.305717] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.314194] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523f169a-1b94-082c-c8d3-340b88b46677, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.476928] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.849s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.479294] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.193s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.498881] env[63293]: INFO nova.scheduler.client.report [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleted allocations for instance 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a [ 962.645794] env[63293]: DEBUG nova.compute.manager [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 962.657414] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327995, 'name': Rename_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.668692] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:32:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.668919] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.669092] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.669317] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.669425] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.669571] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.669794] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.669973] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.670247] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.670433] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.670491] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.678162] env[63293]: DEBUG nova.virt.hardware [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.678417] env[63293]: DEBUG nova.virt.hardware [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.678549] env[63293]: DEBUG nova.virt.hardware [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.678734] env[63293]: DEBUG nova.virt.hardware [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.678897] env[63293]: DEBUG nova.virt.hardware [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.679079] env[63293]: DEBUG nova.virt.hardware [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.679280] env[63293]: DEBUG nova.virt.hardware [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.679460] env[63293]: DEBUG nova.virt.hardware [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.679681] env[63293]: DEBUG nova.virt.hardware [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.679800] env[63293]: DEBUG nova.virt.hardware [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.680028] env[63293]: DEBUG nova.virt.hardware [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.680634] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0fa4efe4-8b35-45a7-8ece-3a50d3d1ab5c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.691186] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf2b067-7df7-4aec-b141-3f1bd4e5c6bf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.701788] env[63293]: DEBUG oslo_concurrency.lockutils [None req-db89a249-52ad-4a56-829c-76b4fe26b3c3 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "f9ccb216-1435-44c5-ab34-a6388d794551" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.617s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.703999] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe953d4-458c-4a56-b014-661055652207 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.707746] env[63293]: DEBUG oslo_vmware.api [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 962.707746] env[63293]: value = "task-1327996" [ 962.707746] env[63293]: _type = "Task" [ 962.707746] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.724393] env[63293]: DEBUG oslo_vmware.api [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327996, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.817818] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523f169a-1b94-082c-c8d3-340b88b46677, 'name': SearchDatastore_Task, 'duration_secs': 0.01322} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.818295] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.818749] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.818949] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.819097] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.819258] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.819703] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8519fc3e-c2c1-44a3-aed8-b007d8336683 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.829336] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.829558] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 962.830379] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03396927-7bd7-4158-8b65-4f5708593b76 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.837873] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 962.837873] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524f26ad-efb6-905e-d214-09f73bbad057" [ 962.837873] env[63293]: _type = "Task" [ 962.837873] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.847837] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524f26ad-efb6-905e-d214-09f73bbad057, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.984508] env[63293]: INFO nova.compute.claims [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 963.006719] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7d3644c5-fdf2-4125-a4a6-b410206a050d tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "5f01ce51-1f5d-44ad-97f0-2306cbf55b9a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.033s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.159420] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327995, 'name': Rename_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.191909] env[63293]: DEBUG nova.compute.manager [req-14115a17-98af-4e2d-afaa-1ef6f4d383d3 req-858c36ed-742b-4af7-9ebb-1484587dd54b service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Received event network-vif-plugged-bda9acec-71cb-4de2-859e-75837898a336 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.192121] env[63293]: DEBUG oslo_concurrency.lockutils [req-14115a17-98af-4e2d-afaa-1ef6f4d383d3 req-858c36ed-742b-4af7-9ebb-1484587dd54b service nova] Acquiring lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.192422] env[63293]: DEBUG oslo_concurrency.lockutils [req-14115a17-98af-4e2d-afaa-1ef6f4d383d3 req-858c36ed-742b-4af7-9ebb-1484587dd54b service nova] Lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.192634] env[63293]: DEBUG oslo_concurrency.lockutils [req-14115a17-98af-4e2d-afaa-1ef6f4d383d3 req-858c36ed-742b-4af7-9ebb-1484587dd54b service nova] Lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.192816] env[63293]: DEBUG nova.compute.manager [req-14115a17-98af-4e2d-afaa-1ef6f4d383d3 req-858c36ed-742b-4af7-9ebb-1484587dd54b service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] No waiting events found dispatching network-vif-plugged-bda9acec-71cb-4de2-859e-75837898a336 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.193140] env[63293]: WARNING nova.compute.manager [req-14115a17-98af-4e2d-afaa-1ef6f4d383d3 req-858c36ed-742b-4af7-9ebb-1484587dd54b service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Received unexpected event network-vif-plugged-bda9acec-71cb-4de2-859e-75837898a336 for instance with vm_state building and task_state spawning. [ 963.219077] env[63293]: DEBUG oslo_vmware.api [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327996, 'name': ReconfigVM_Task, 'duration_secs': 0.167368} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.219077] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance '9ddf4f23-1279-4cbd-8212-10f344060445' progress to 33 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 963.349022] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524f26ad-efb6-905e-d214-09f73bbad057, 'name': SearchDatastore_Task, 'duration_secs': 0.010808} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.349842] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec2fc9f5-fbb1-43bc-a5a3-d55a9a62a019 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.356123] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 963.356123] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e0d3c5-eb1c-6ff2-a548-d0583517c999" [ 963.356123] env[63293]: _type = "Task" [ 963.356123] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.364730] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e0d3c5-eb1c-6ff2-a548-d0583517c999, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.392942] env[63293]: DEBUG nova.network.neutron [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Successfully updated port: bda9acec-71cb-4de2-859e-75837898a336 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.493393] env[63293]: INFO nova.compute.resource_tracker [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating resource usage from migration a719d2d5-2903-4e76-9d8c-cc2588bcda26 [ 963.659170] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327995, 'name': Rename_Task, 'duration_secs': 1.271149} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.659491] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.659745] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65de4c1d-63e2-46f1-9420-403e1a8bd7d4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.666866] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 963.666866] env[63293]: value = "task-1327997" [ 963.666866] env[63293]: _type = "Task" [ 963.666866] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.677339] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327997, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.727233] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 963.727233] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 963.727233] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.727233] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 963.727233] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.727233] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 963.727233] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 963.727233] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 963.727923] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 963.728232] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 963.728555] env[63293]: DEBUG nova.virt.hardware [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 963.735707] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Reconfiguring VM instance instance-00000045 to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 963.738970] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29238e96-2667-4be7-835f-a75e902f3145 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.763021] env[63293]: DEBUG oslo_vmware.api [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 963.763021] env[63293]: value = "task-1327998" [ 963.763021] env[63293]: _type = "Task" [ 963.763021] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.773291] env[63293]: DEBUG oslo_vmware.api [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327998, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.781133] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b07f78-09e4-4b13-b4a4-9d86c48751b5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.788432] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5186a3-99d5-49b0-aedc-c8572426eb7d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.820912] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c085b1b-9b1f-4716-be45-b77106bfce8b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.829668] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c7190f-9cbb-4479-ac42-66fb297a81d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.847742] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 963.849054] env[63293]: DEBUG nova.compute.provider_tree [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.850957] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 963.872550] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e0d3c5-eb1c-6ff2-a548-d0583517c999, 'name': SearchDatastore_Task, 'duration_secs': 0.016387} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.872550] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.872550] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 657d5f21-b985-4e3c-8799-b3aae2b1c509/657d5f21-b985-4e3c-8799-b3aae2b1c509.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 963.872550] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0029ed23-37d7-4c97-83ca-1d00bef72905 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.879428] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 963.879428] env[63293]: value = "task-1327999" [ 963.879428] env[63293]: _type = "Task" [ 963.879428] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.891377] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327999, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.897895] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "refresh_cache-9d3fcf27-2c2c-4975-a802-a5c0859640d0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.897895] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "refresh_cache-9d3fcf27-2c2c-4975-a802-a5c0859640d0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.897895] env[63293]: DEBUG nova.network.neutron [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.181906] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327997, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.273602] env[63293]: DEBUG oslo_vmware.api [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1327998, 'name': ReconfigVM_Task, 'duration_secs': 0.175668} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.273893] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Reconfigured VM instance instance-00000045 to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 964.274621] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04fd8303-937d-43b9-8ce5-8ca4ff54d8b8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.303640] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 9ddf4f23-1279-4cbd-8212-10f344060445/9ddf4f23-1279-4cbd-8212-10f344060445.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 964.304558] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd5c4695-17fe-47be-b8e8-01e8f3a8f0d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.327143] env[63293]: DEBUG oslo_vmware.api [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 964.327143] env[63293]: value = "task-1328000" [ 964.327143] env[63293]: _type = "Task" [ 964.327143] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.337959] env[63293]: DEBUG oslo_vmware.api [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328000, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.354993] env[63293]: DEBUG nova.scheduler.client.report [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.367182] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 964.368141] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Starting heal instance info cache {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 964.391729] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327999, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.442659] env[63293]: DEBUG nova.network.neutron [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 964.684095] env[63293]: DEBUG oslo_vmware.api [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1327997, 'name': PowerOnVM_Task, 'duration_secs': 0.523422} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.684312] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.684714] env[63293]: INFO nova.compute.manager [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Took 9.20 seconds to spawn the instance on the hypervisor. [ 964.684714] env[63293]: DEBUG nova.compute.manager [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.685895] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c422be14-2fd7-4c72-882a-6ef855a977cd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.764602] env[63293]: DEBUG nova.network.neutron [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Updating instance_info_cache with network_info: [{"id": "bda9acec-71cb-4de2-859e-75837898a336", "address": "fa:16:3e:ea:f2:1e", "network": {"id": "340bfd3b-7dff-454b-87df-788003988adc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-388838680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2af826d819b4cc4ab27ff8b95144039", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbda9acec-71", "ovs_interfaceid": "bda9acec-71cb-4de2-859e-75837898a336", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.837354] env[63293]: DEBUG oslo_vmware.api [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328000, 'name': ReconfigVM_Task, 'duration_secs': 0.323093} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.837654] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 9ddf4f23-1279-4cbd-8212-10f344060445/9ddf4f23-1279-4cbd-8212-10f344060445.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.837926] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance '9ddf4f23-1279-4cbd-8212-10f344060445' progress to 50 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 964.867896] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.389s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.868141] env[63293]: INFO nova.compute.manager [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Migrating [ 964.876595] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.824s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.876825] env[63293]: DEBUG nova.objects.instance [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lazy-loading 'resources' on Instance uuid 6f1fb925-72c2-4309-b3fd-aea217d38bf2 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.896914] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1327999, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.593111} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.897323] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 657d5f21-b985-4e3c-8799-b3aae2b1c509/657d5f21-b985-4e3c-8799-b3aae2b1c509.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 964.897478] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 964.898150] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef6d8377-5339-46b3-ae4c-c4b2b24987b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.906591] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 964.906591] env[63293]: value = "task-1328001" [ 964.906591] env[63293]: _type = "Task" [ 964.906591] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.918846] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328001, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.212168] env[63293]: INFO nova.compute.manager [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Took 21.89 seconds to build instance. [ 965.239324] env[63293]: DEBUG nova.compute.manager [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Received event network-changed-bda9acec-71cb-4de2-859e-75837898a336 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.239523] env[63293]: DEBUG nova.compute.manager [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Refreshing instance network info cache due to event network-changed-bda9acec-71cb-4de2-859e-75837898a336. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.239714] env[63293]: DEBUG oslo_concurrency.lockutils [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] Acquiring lock "refresh_cache-9d3fcf27-2c2c-4975-a802-a5c0859640d0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.267116] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "refresh_cache-9d3fcf27-2c2c-4975-a802-a5c0859640d0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.267423] env[63293]: DEBUG nova.compute.manager [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Instance network_info: |[{"id": "bda9acec-71cb-4de2-859e-75837898a336", "address": "fa:16:3e:ea:f2:1e", "network": {"id": "340bfd3b-7dff-454b-87df-788003988adc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-388838680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2af826d819b4cc4ab27ff8b95144039", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbda9acec-71", "ovs_interfaceid": "bda9acec-71cb-4de2-859e-75837898a336", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.267737] env[63293]: DEBUG oslo_concurrency.lockutils [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] Acquired lock "refresh_cache-9d3fcf27-2c2c-4975-a802-a5c0859640d0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.267921] env[63293]: DEBUG nova.network.neutron [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Refreshing network info cache for port bda9acec-71cb-4de2-859e-75837898a336 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.269173] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:f2:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb2eaecd-9701-4504-9fcb-fb1a420ead72', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bda9acec-71cb-4de2-859e-75837898a336', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.281449] env[63293]: DEBUG oslo.service.loopingcall [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.281564] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 965.281772] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-974ff085-1780-409c-8805-43cbd8290da0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.304219] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.304219] env[63293]: value = "task-1328002" [ 965.304219] env[63293]: _type = "Task" [ 965.304219] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.314468] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328002, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.348477] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af07fb2-e6b2-4bcb-ae85-a541f09f80ae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.371430] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c7d01b-3ea0-4c94-9c0c-d108eb4cf2b8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.396527] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.396759] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.396963] env[63293]: DEBUG nova.network.neutron [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 965.398365] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance '9ddf4f23-1279-4cbd-8212-10f344060445' progress to 67 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 965.418645] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328001, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078948} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.419165] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.420940] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf19f7e-b34b-4d88-8a07-3f2398a1c1a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.454048] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 657d5f21-b985-4e3c-8799-b3aae2b1c509/657d5f21-b985-4e3c-8799-b3aae2b1c509.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.458646] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c7acd91-b6c6-4af7-9e37-5a46faae82ee {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.482485] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 965.482485] env[63293]: value = "task-1328003" [ 965.482485] env[63293]: _type = "Task" [ 965.482485] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.495665] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328003, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.669509] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66bd6730-823d-449d-8dfd-c075f12acfa2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.678135] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639c0dbe-91a3-49d2-82b8-cd89a605dafb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.711778] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1602887c-1228-4962-9e57-f646bea2b985 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.715336] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cef109e3-edfb-437a-82b9-df1456abe979 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "66be1abd-c541-4b8a-8d72-98bde03b1888" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.400s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.722532] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da55b67-2295-4a31-ad4a-e7c7e88007a2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.737997] env[63293]: DEBUG nova.compute.provider_tree [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.815668] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328002, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.914293] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290d085e-f9d9-40f0-ac4e-181e8154f81a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.924357] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-07ee246d-bcd7-46b9-a9c3-4034cc91ad75 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Suspending the VM {{(pid=63293) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 965.924620] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-e6df7b2e-5fcf-4451-9c53-102b5668e578 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.931573] env[63293]: DEBUG oslo_vmware.api [None req-07ee246d-bcd7-46b9-a9c3-4034cc91ad75 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 965.931573] env[63293]: value = "task-1328004" [ 965.931573] env[63293]: _type = "Task" [ 965.931573] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.941081] env[63293]: DEBUG oslo_vmware.api [None req-07ee246d-bcd7-46b9-a9c3-4034cc91ad75 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328004, 'name': SuspendVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.951909] env[63293]: DEBUG nova.network.neutron [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Port 4dbccf6a-053e-4933-91bc-1b7af5a260be binding to destination host cpu-1 is already ACTIVE {{(pid=63293) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 965.993771] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328003, 'name': ReconfigVM_Task, 'duration_secs': 0.276619} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.994092] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 657d5f21-b985-4e3c-8799-b3aae2b1c509/657d5f21-b985-4e3c-8799-b3aae2b1c509.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.994857] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-664314d4-bc23-40f7-866f-92514bbb7cb4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.001887] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 966.001887] env[63293]: value = "task-1328005" [ 966.001887] env[63293]: _type = "Task" [ 966.001887] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.013298] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328005, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.159831] env[63293]: DEBUG nova.network.neutron [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Updated VIF entry in instance network info cache for port bda9acec-71cb-4de2-859e-75837898a336. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.160323] env[63293]: DEBUG nova.network.neutron [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Updating instance_info_cache with network_info: [{"id": "bda9acec-71cb-4de2-859e-75837898a336", "address": "fa:16:3e:ea:f2:1e", "network": {"id": "340bfd3b-7dff-454b-87df-788003988adc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-388838680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2af826d819b4cc4ab27ff8b95144039", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbda9acec-71", "ovs_interfaceid": "bda9acec-71cb-4de2-859e-75837898a336", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.217185] env[63293]: DEBUG nova.network.neutron [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance_info_cache with network_info: [{"id": "a20688c6-3fbe-4355-85f9-334019e37f43", "address": "fa:16:3e:08:30:f7", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa20688c6-3f", "ovs_interfaceid": "a20688c6-3fbe-4355-85f9-334019e37f43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.241057] env[63293]: DEBUG nova.scheduler.client.report [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.317227] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328002, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.443630] env[63293]: DEBUG oslo_vmware.api [None req-07ee246d-bcd7-46b9-a9c3-4034cc91ad75 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328004, 'name': SuspendVM_Task} progress is 62%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.513026] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328005, 'name': Rename_Task, 'duration_secs': 0.152356} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.513373] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 966.513628] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1867ad86-1a55-4a7f-bc7c-fb9fe7958850 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.520576] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 966.520576] env[63293]: value = "task-1328006" [ 966.520576] env[63293]: _type = "Task" [ 966.520576] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.528816] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328006, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.663779] env[63293]: DEBUG oslo_concurrency.lockutils [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] Releasing lock "refresh_cache-9d3fcf27-2c2c-4975-a802-a5c0859640d0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.664332] env[63293]: DEBUG nova.compute.manager [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Received event network-changed-df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.664449] env[63293]: DEBUG nova.compute.manager [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing instance network info cache due to event network-changed-df1c83c6-23d7-439b-a496-234c48738d81. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 966.664635] env[63293]: DEBUG oslo_concurrency.lockutils [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] Acquiring lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.664822] env[63293]: DEBUG oslo_concurrency.lockutils [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] Acquired lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.665045] env[63293]: DEBUG nova.network.neutron [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing network info cache for port df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 966.720470] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.748181] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.871s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.750897] env[63293]: DEBUG oslo_concurrency.lockutils [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.470s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.751160] env[63293]: DEBUG nova.objects.instance [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'resources' on Instance uuid 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.770285] env[63293]: INFO nova.scheduler.client.report [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Deleted allocations for instance 6f1fb925-72c2-4309-b3fd-aea217d38bf2 [ 966.817241] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328002, 'name': CreateVM_Task, 'duration_secs': 1.299354} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.818027] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 966.818761] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.818938] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.819285] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.819827] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b6520f0-09c0-4781-8385-66c5bc844154 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.825724] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 966.825724] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527066fc-8df4-d8ef-52ab-888b65b99fac" [ 966.825724] env[63293]: _type = "Task" [ 966.825724] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.836332] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527066fc-8df4-d8ef-52ab-888b65b99fac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.947915] env[63293]: DEBUG oslo_vmware.api [None req-07ee246d-bcd7-46b9-a9c3-4034cc91ad75 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328004, 'name': SuspendVM_Task} progress is 62%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.961477] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.961624] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquired lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.961772] env[63293]: DEBUG nova.network.neutron [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Forcefully refreshing network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 966.986848] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "9ddf4f23-1279-4cbd-8212-10f344060445-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.987108] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.987300] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.033244] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328006, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.279611] env[63293]: DEBUG nova.compute.manager [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Received event network-changed-5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.279884] env[63293]: DEBUG nova.compute.manager [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Refreshing instance network info cache due to event network-changed-5d29fa7c-6329-4700-b988-ac08592d4df7. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 967.280067] env[63293]: DEBUG oslo_concurrency.lockutils [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] Acquiring lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.280225] env[63293]: DEBUG oslo_concurrency.lockutils [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] Acquired lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.280384] env[63293]: DEBUG nova.network.neutron [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Refreshing network info cache for port 5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.281647] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c3b4ae91-fa5e-4488-815f-2de19137f68b tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "6f1fb925-72c2-4309-b3fd-aea217d38bf2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.101s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.339471] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527066fc-8df4-d8ef-52ab-888b65b99fac, 'name': SearchDatastore_Task, 'duration_secs': 0.011365} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.339471] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.339471] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.339616] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.339756] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.339970] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.340706] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-934f4e03-5879-4b13-862f-7590c04ac482 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.351332] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.351332] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 967.352088] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4272f44c-d6bd-4fbc-aa8f-9cf015b03116 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.360847] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 967.360847] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523a817c-0ea3-ca11-50f3-ac9b6152213e" [ 967.360847] env[63293]: _type = "Task" [ 967.360847] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.375754] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523a817c-0ea3-ca11-50f3-ac9b6152213e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.445644] env[63293]: DEBUG oslo_vmware.api [None req-07ee246d-bcd7-46b9-a9c3-4034cc91ad75 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328004, 'name': SuspendVM_Task, 'duration_secs': 1.228024} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.445923] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-07ee246d-bcd7-46b9-a9c3-4034cc91ad75 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Suspended the VM {{(pid=63293) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 967.446152] env[63293]: DEBUG nova.compute.manager [None req-07ee246d-bcd7-46b9-a9c3-4034cc91ad75 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.448737] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d77816-52bf-4d3f-9ad6-0ecc5fb3aae7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.488148] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5158263-9283-4c33-a8d3-d616412a1207 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.497807] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8c0e35-c79c-4474-9b45-e8932f35aa06 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.502853] env[63293]: DEBUG nova.network.neutron [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updated VIF entry in instance network info cache for port df1c83c6-23d7-439b-a496-234c48738d81. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 967.502853] env[63293]: DEBUG nova.network.neutron [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updating instance_info_cache with network_info: [{"id": "df1c83c6-23d7-439b-a496-234c48738d81", "address": "fa:16:3e:0f:9f:14", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1c83c6-23", "ovs_interfaceid": "df1c83c6-23d7-439b-a496-234c48738d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.538665] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccbbccee-a17c-4a11-be95-68e2d95f0cff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.552892] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbaeb91-0fa8-4807-8d9a-70c870d4eda8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.557128] env[63293]: DEBUG oslo_vmware.api [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328006, 'name': PowerOnVM_Task, 'duration_secs': 0.730482} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.557647] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 967.557874] env[63293]: INFO nova.compute.manager [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Took 9.72 seconds to spawn the instance on the hypervisor. [ 967.558067] env[63293]: DEBUG nova.compute.manager [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.559173] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3128a336-f999-4444-b8f4-85641295f041 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.570046] env[63293]: DEBUG nova.compute.provider_tree [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.789320] env[63293]: DEBUG nova.network.neutron [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updating instance_info_cache with network_info: [{"id": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "address": "fa:16:3e:c0:e8:15", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0576b708-1b", "ovs_interfaceid": "0576b708-1b9b-4159-8079-9b97d8b4e70d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.876116] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523a817c-0ea3-ca11-50f3-ac9b6152213e, 'name': SearchDatastore_Task, 'duration_secs': 0.010751} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.876936] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24efac27-6baa-4005-bb0f-0f14cee0d959 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.882836] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 967.882836] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52deb25f-3c48-4292-295a-6a52798a7a87" [ 967.882836] env[63293]: _type = "Task" [ 967.882836] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.890526] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52deb25f-3c48-4292-295a-6a52798a7a87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.007275] env[63293]: DEBUG oslo_concurrency.lockutils [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] Releasing lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.007275] env[63293]: DEBUG nova.compute.manager [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Received event network-changed-5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.007345] env[63293]: DEBUG nova.compute.manager [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Refreshing instance network info cache due to event network-changed-5d29fa7c-6329-4700-b988-ac08592d4df7. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 968.007531] env[63293]: DEBUG oslo_concurrency.lockutils [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] Acquiring lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.035806] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.035992] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.036192] env[63293]: DEBUG nova.network.neutron [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.062197] env[63293]: DEBUG nova.network.neutron [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updated VIF entry in instance network info cache for port 5d29fa7c-6329-4700-b988-ac08592d4df7. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 968.062534] env[63293]: DEBUG nova.network.neutron [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updating instance_info_cache with network_info: [{"id": "5d29fa7c-6329-4700-b988-ac08592d4df7", "address": "fa:16:3e:56:b9:c4", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d29fa7c-63", "ovs_interfaceid": "5d29fa7c-6329-4700-b988-ac08592d4df7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.072777] env[63293]: DEBUG nova.scheduler.client.report [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.084155] env[63293]: INFO nova.compute.manager [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Took 22.57 seconds to build instance. [ 968.240024] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3f2d8f-8fce-4899-8af4-b383e1d520f6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.258584] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance 'f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a' progress to 0 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 968.292095] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Releasing lock "refresh_cache-ad585ebb-2072-45df-b645-94c9fa93576b" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.292095] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updated the network info_cache for instance {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 968.292095] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 968.292350] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 968.292542] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 968.292728] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 968.292913] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 968.293330] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 968.293483] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63293) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 968.293653] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 968.334272] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "66be1abd-c541-4b8a-8d72-98bde03b1888" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.334550] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "66be1abd-c541-4b8a-8d72-98bde03b1888" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.334766] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "66be1abd-c541-4b8a-8d72-98bde03b1888-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.334970] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "66be1abd-c541-4b8a-8d72-98bde03b1888-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.335166] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "66be1abd-c541-4b8a-8d72-98bde03b1888-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.337523] env[63293]: INFO nova.compute.manager [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Terminating instance [ 968.339285] env[63293]: DEBUG nova.compute.manager [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.339491] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.340394] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764e502d-6bd8-4e2a-852a-750cdeec3ee3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.348604] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 968.348857] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9cadda55-a48c-414a-9fff-60e0e31316db {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.394030] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52deb25f-3c48-4292-295a-6a52798a7a87, 'name': SearchDatastore_Task, 'duration_secs': 0.013738} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.394185] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.394328] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 9d3fcf27-2c2c-4975-a802-a5c0859640d0/9d3fcf27-2c2c-4975-a802-a5c0859640d0.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 968.394587] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c06caeb-73a6-48e3-a27e-6126f8720dfd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.401830] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 968.401830] env[63293]: value = "task-1328008" [ 968.401830] env[63293]: _type = "Task" [ 968.401830] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.410985] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.425621] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 968.425848] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 968.426070] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleting the datastore file [datastore2] 66be1abd-c541-4b8a-8d72-98bde03b1888 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.426384] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6ec562a-b29f-43b7-a93d-df6ab40d1faf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.433713] env[63293]: DEBUG oslo_vmware.api [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 968.433713] env[63293]: value = "task-1328009" [ 968.433713] env[63293]: _type = "Task" [ 968.433713] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.442757] env[63293]: DEBUG oslo_vmware.api [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328009, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.564769] env[63293]: DEBUG oslo_concurrency.lockutils [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] Releasing lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.565250] env[63293]: DEBUG nova.compute.manager [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Received event network-changed-df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.565563] env[63293]: DEBUG nova.compute.manager [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing instance network info cache due to event network-changed-df1c83c6-23d7-439b-a496-234c48738d81. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 968.565882] env[63293]: DEBUG oslo_concurrency.lockutils [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] Acquiring lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.566580] env[63293]: DEBUG oslo_concurrency.lockutils [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] Acquired lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.566580] env[63293]: DEBUG nova.network.neutron [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing network info cache for port df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 968.570182] env[63293]: DEBUG oslo_concurrency.lockutils [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] Acquired lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.570270] env[63293]: DEBUG nova.network.neutron [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Refreshing network info cache for port 5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 968.577918] env[63293]: DEBUG oslo_concurrency.lockutils [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.585689] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9abbbef2-bc55-495d-bff7-fe3db1de3bd7 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "657d5f21-b985-4e3c-8799-b3aae2b1c509" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.083s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.600996] env[63293]: INFO nova.scheduler.client.report [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Deleted allocations for instance 0977c79a-2ef3-4ec9-a0a0-de1ed5799723 [ 968.764407] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.764773] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-edd73896-046a-4bdf-8a08-b48c0a252a49 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.774026] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 968.774026] env[63293]: value = "task-1328010" [ 968.774026] env[63293]: _type = "Task" [ 968.774026] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.786910] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328010, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.797150] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.797453] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.797558] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.797712] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63293) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 968.799054] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2c5528-da17-4603-bb52-ab62f6d6b088 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.814867] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af42681-0e19-4055-989a-ad3d6b4f71f3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.842790] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac45f4f-cebe-4f7d-adfc-971a4117012d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.851311] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a194c7db-d1c8-4900-81c7-15e34fea9950 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.876499] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180208MB free_disk=132GB free_vcpus=48 pci_devices=None {{(pid=63293) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 968.876651] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.876811] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.880364] env[63293]: DEBUG nova.network.neutron [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance_info_cache with network_info: [{"id": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "address": "fa:16:3e:93:d3:25", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dbccf6a-05", "ovs_interfaceid": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.913988] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328008, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.946347] env[63293]: DEBUG oslo_vmware.api [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328009, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247743} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.946987] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.947232] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 968.947435] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 968.947616] env[63293]: INFO nova.compute.manager [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Took 0.61 seconds to destroy the instance on the hypervisor. [ 968.947885] env[63293]: DEBUG oslo.service.loopingcall [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.948106] env[63293]: DEBUG nova.compute.manager [-] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 968.948208] env[63293]: DEBUG nova.network.neutron [-] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 969.110148] env[63293]: DEBUG oslo_concurrency.lockutils [None req-05f79f02-b65d-48dc-9741-5c2d6ae35a54 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "0977c79a-2ef3-4ec9-a0a0-de1ed5799723" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.178s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.285309] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328010, 'name': PowerOffVM_Task, 'duration_secs': 0.273844} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.287937] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.288250] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance 'f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a' progress to 17 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 969.383511] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.400808] env[63293]: DEBUG nova.network.neutron [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updated VIF entry in instance network info cache for port 5d29fa7c-6329-4700-b988-ac08592d4df7. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 969.400808] env[63293]: DEBUG nova.network.neutron [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updating instance_info_cache with network_info: [{"id": "5d29fa7c-6329-4700-b988-ac08592d4df7", "address": "fa:16:3e:56:b9:c4", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d29fa7c-63", "ovs_interfaceid": "5d29fa7c-6329-4700-b988-ac08592d4df7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.412781] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328008, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.498691] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "657d5f21-b985-4e3c-8799-b3aae2b1c509" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.499046] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "657d5f21-b985-4e3c-8799-b3aae2b1c509" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.499291] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "657d5f21-b985-4e3c-8799-b3aae2b1c509-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.502335] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "657d5f21-b985-4e3c-8799-b3aae2b1c509-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.502335] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "657d5f21-b985-4e3c-8799-b3aae2b1c509-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.502335] env[63293]: INFO nova.compute.manager [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Terminating instance [ 969.503538] env[63293]: DEBUG nova.compute.manager [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.503612] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.504412] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b496086d-8d79-4575-9574-05be9e08e656 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.514425] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.515035] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-174a292b-b896-4a18-9682-bee9927a42fb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.522033] env[63293]: DEBUG oslo_vmware.api [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 969.522033] env[63293]: value = "task-1328011" [ 969.522033] env[63293]: _type = "Task" [ 969.522033] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.527316] env[63293]: DEBUG nova.compute.manager [req-7ab0ecbd-4201-44a8-88d0-828b641dcb85 req-1984c4a6-f1dd-48d0-a43d-cca5483245e3 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Received event network-vif-deleted-3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.527568] env[63293]: INFO nova.compute.manager [req-7ab0ecbd-4201-44a8-88d0-828b641dcb85 req-1984c4a6-f1dd-48d0-a43d-cca5483245e3 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Neutron deleted interface 3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf; detaching it from the instance and deleting it from the info cache [ 969.527864] env[63293]: DEBUG nova.network.neutron [req-7ab0ecbd-4201-44a8-88d0-828b641dcb85 req-1984c4a6-f1dd-48d0-a43d-cca5483245e3 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.535826] env[63293]: DEBUG oslo_vmware.api [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328011, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.536949] env[63293]: DEBUG nova.network.neutron [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updated VIF entry in instance network info cache for port df1c83c6-23d7-439b-a496-234c48738d81. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 969.539188] env[63293]: DEBUG nova.network.neutron [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updating instance_info_cache with network_info: [{"id": "df1c83c6-23d7-439b-a496-234c48738d81", "address": "fa:16:3e:0f:9f:14", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1c83c6-23", "ovs_interfaceid": "df1c83c6-23d7-439b-a496-234c48738d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.795359] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:32:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 969.795647] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 969.795808] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 969.795996] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 969.796167] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 969.796324] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 969.796531] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 969.796697] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 969.796925] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 969.797043] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 969.797224] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 969.802778] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-534057a2-6e29-49e9-bd0c-2fd8fe28e122 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.819831] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 969.819831] env[63293]: value = "task-1328012" [ 969.819831] env[63293]: _type = "Task" [ 969.819831] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.829083] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328012, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.904242] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Applying migration context for instance 9ddf4f23-1279-4cbd-8212-10f344060445 as it has an incoming, in-progress migration a66e3649-f4ed-4fb1-8bca-bb060f430e05. Migration status is post-migrating {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 969.904449] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Applying migration context for instance f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a as it has an incoming, in-progress migration a719d2d5-2903-4e76-9d8c-cc2588bcda26. Migration status is migrating {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 969.905905] env[63293]: INFO nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating resource usage from migration a66e3649-f4ed-4fb1-8bca-bb060f430e05 [ 969.906232] env[63293]: INFO nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating resource usage from migration a719d2d5-2903-4e76-9d8c-cc2588bcda26 [ 969.911590] env[63293]: DEBUG oslo_concurrency.lockutils [req-8a3055c9-d893-47a1-a8d6-226f92913011 req-01b61158-e221-4ba4-8647-a184f8d91cfb service nova] Releasing lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.921975] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328008, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.924971] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14867f55-7750-4a2e-93a6-f3a68421a4a2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.928116] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance d753598d-a92a-4515-9ad1-d386294f7a99 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.928192] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance ad585ebb-2072-45df-b645-94c9fa93576b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.929041] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance f429869f-9e9f-443e-a980-67a80c3bd799 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.929041] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 35cbf277-b85a-4101-beae-0125fb5a61f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.929041] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.929041] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 828180c0-b7f8-4666-9f5d-e6741ef6495a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.929041] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance f9ccb216-1435-44c5-ab34-a6388d794551 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.929041] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 66be1abd-c541-4b8a-8d72-98bde03b1888 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.929041] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 657d5f21-b985-4e3c-8799-b3aae2b1c509 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.929292] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Migration a66e3649-f4ed-4fb1-8bca-bb060f430e05 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 969.929292] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 9ddf4f23-1279-4cbd-8212-10f344060445 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.929292] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 9d3fcf27-2c2c-4975-a802-a5c0859640d0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.929375] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Migration a719d2d5-2903-4e76-9d8c-cc2588bcda26 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 969.929483] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.929689] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 969.929822] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3328MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 969.953700] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77adc48e-e405-4a4d-9390-2175471fc651 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.962063] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance '9ddf4f23-1279-4cbd-8212-10f344060445' progress to 83 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 969.993425] env[63293]: DEBUG nova.network.neutron [-] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.033876] env[63293]: DEBUG oslo_vmware.api [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328011, 'name': PowerOffVM_Task, 'duration_secs': 0.238097} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.034179] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9959a92e-36b6-4417-ad2a-19a02f6d3581 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.035994] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.036185] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.036426] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9b57019-fa45-4961-b441-3a88fcb1a68e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.042881] env[63293]: DEBUG oslo_concurrency.lockutils [req-62c3cc39-7535-41ba-bb1a-f68f12fe65c1 req-661fc425-078a-4040-ac58-6426d741fe5a service nova] Releasing lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.049747] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae552c10-4529-414d-91fd-68e0acae0419 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.084578] env[63293]: DEBUG nova.compute.manager [req-7ab0ecbd-4201-44a8-88d0-828b641dcb85 req-1984c4a6-f1dd-48d0-a43d-cca5483245e3 service nova] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Detach interface failed, port_id=3fccaa93-b995-4a2b-aac1-0b8b3d8cd2bf, reason: Instance 66be1abd-c541-4b8a-8d72-98bde03b1888 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 970.129555] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.130226] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.130302] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleting the datastore file [datastore1] 657d5f21-b985-4e3c-8799-b3aae2b1c509 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.130624] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1c48fc35-964e-4ec3-9bbe-cabef4e2a70c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.140607] env[63293]: DEBUG oslo_vmware.api [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 970.140607] env[63293]: value = "task-1328014" [ 970.140607] env[63293]: _type = "Task" [ 970.140607] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.151066] env[63293]: DEBUG oslo_vmware.api [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.164629] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057a6bbe-0670-4873-bbea-9206e3dba7a0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.173472] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf329e2-1fa9-4f92-9f1a-8cba14571954 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.206975] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfa61bb-dad3-4ddb-ba6c-5fc1f4c8cdbd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.216262] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e56755d-5a81-4e05-b32d-81f8f61d48a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.231314] env[63293]: DEBUG nova.compute.provider_tree [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.331517] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328012, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.422527] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328008, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.68502} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.422788] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 9d3fcf27-2c2c-4975-a802-a5c0859640d0/9d3fcf27-2c2c-4975-a802-a5c0859640d0.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 970.423026] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 970.423369] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-537fe890-7e0f-4ee9-a86e-65a0a0a88f0e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.430831] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 970.430831] env[63293]: value = "task-1328015" [ 970.430831] env[63293]: _type = "Task" [ 970.430831] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.440540] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328015, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.469337] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-1506a8d1-d86e-4232-9e44-90d7435d4bea tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance '9ddf4f23-1279-4cbd-8212-10f344060445' progress to 100 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 970.495993] env[63293]: INFO nova.compute.manager [-] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Took 1.55 seconds to deallocate network for instance. [ 970.652960] env[63293]: DEBUG oslo_vmware.api [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.734689] env[63293]: DEBUG nova.scheduler.client.report [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.831475] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328012, 'name': ReconfigVM_Task, 'duration_secs': 0.743468} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.831797] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance 'f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a' progress to 33 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 970.941934] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328015, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.128478} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.942135] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.942866] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d795e007-8655-42de-a587-a81f9096fe87 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.966169] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 9d3fcf27-2c2c-4975-a802-a5c0859640d0/9d3fcf27-2c2c-4975-a802-a5c0859640d0.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.966464] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37e3e0c1-3f6f-4730-ae43-690bd51b034b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.990020] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 970.990020] env[63293]: value = "task-1328016" [ 970.990020] env[63293]: _type = "Task" [ 970.990020] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.999666] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328016, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.003041] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.152517] env[63293]: DEBUG oslo_vmware.api [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328014, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.631919} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.152866] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.153564] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 971.154306] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 971.154306] env[63293]: INFO nova.compute.manager [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Took 1.65 seconds to destroy the instance on the hypervisor. [ 971.154306] env[63293]: DEBUG oslo.service.loopingcall [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.154520] env[63293]: DEBUG nova.compute.manager [-] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 971.154647] env[63293]: DEBUG nova.network.neutron [-] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 971.239304] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 971.239512] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.363s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.239792] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.237s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.240056] env[63293]: DEBUG nova.objects.instance [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lazy-loading 'resources' on Instance uuid 66be1abd-c541-4b8a-8d72-98bde03b1888 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.338790] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.339106] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.339275] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.339463] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.339613] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.339764] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.339983] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.340158] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.340394] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.340578] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.340803] env[63293]: DEBUG nova.virt.hardware [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.346422] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Reconfiguring VM instance instance-0000005a to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 971.346782] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-061c21e0-6a34-407d-ae7c-83dad00935d6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.369838] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 971.369838] env[63293]: value = "task-1328017" [ 971.369838] env[63293]: _type = "Task" [ 971.369838] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.384019] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328017, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.500446] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328016, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.554135] env[63293]: DEBUG nova.compute.manager [req-b3ea6f7a-b205-454e-b121-8c572e90360f req-5a658d29-1f81-405d-a89d-214d54e8df53 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Received event network-vif-deleted-0a09da68-14b4-4caa-9967-6b00717e08d7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.554135] env[63293]: INFO nova.compute.manager [req-b3ea6f7a-b205-454e-b121-8c572e90360f req-5a658d29-1f81-405d-a89d-214d54e8df53 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Neutron deleted interface 0a09da68-14b4-4caa-9967-6b00717e08d7; detaching it from the instance and deleting it from the info cache [ 971.554135] env[63293]: DEBUG nova.network.neutron [req-b3ea6f7a-b205-454e-b121-8c572e90360f req-5a658d29-1f81-405d-a89d-214d54e8df53 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.839657] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.839921] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.881319] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328017, 'name': ReconfigVM_Task, 'duration_secs': 0.173421} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.881573] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Reconfigured VM instance instance-0000005a to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 971.882411] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08cd3104-4792-4d9a-86d0-e9709058052f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.899781] env[63293]: DEBUG nova.network.neutron [-] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.908283] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a/f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.909317] env[63293]: INFO nova.compute.manager [-] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Took 0.75 seconds to deallocate network for instance. [ 971.912349] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b9c6d32-5d5e-47fe-830b-2ba78c585d1a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.939512] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 971.939512] env[63293]: value = "task-1328018" [ 971.939512] env[63293]: _type = "Task" [ 971.939512] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.951605] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328018, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.015019] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328016, 'name': ReconfigVM_Task, 'duration_secs': 0.849987} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.015019] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 9d3fcf27-2c2c-4975-a802-a5c0859640d0/9d3fcf27-2c2c-4975-a802-a5c0859640d0.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.015019] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f55157c-4630-4197-88b8-d63f184bd887 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.017950] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69dc6c92-2d18-4026-ac35-f67ab782c4c7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.027164] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d7b73b-5c45-4826-915a-d0683de97e16 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.030616] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 972.030616] env[63293]: value = "task-1328019" [ 972.030616] env[63293]: _type = "Task" [ 972.030616] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.061885] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bbba6746-e25e-44e8-a645-fb37c5df6daf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.064570] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1884b840-dec4-4eb2-81d2-b1d164231214 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.070393] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328019, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.078493] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4f116d-ab78-43a9-9263-01c2e49a2b7f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.083353] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123011fc-6ada-426a-a012-2930dbdea6ff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.110156] env[63293]: DEBUG nova.compute.provider_tree [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.120935] env[63293]: DEBUG nova.compute.manager [req-b3ea6f7a-b205-454e-b121-8c572e90360f req-5a658d29-1f81-405d-a89d-214d54e8df53 service nova] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Detach interface failed, port_id=0a09da68-14b4-4caa-9967-6b00717e08d7, reason: Instance 657d5f21-b985-4e3c-8799-b3aae2b1c509 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 972.302733] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "9ddf4f23-1279-4cbd-8212-10f344060445" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.303100] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.303422] env[63293]: DEBUG nova.compute.manager [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Going to confirm migration 2 {{(pid=63293) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 972.345338] env[63293]: DEBUG nova.compute.manager [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 972.433822] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.449973] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328018, 'name': ReconfigVM_Task, 'duration_secs': 0.290111} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.450299] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Reconfigured VM instance instance-0000005a to attach disk [datastore2] f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a/f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.450590] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance 'f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a' progress to 50 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 972.541383] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328019, 'name': Rename_Task, 'duration_secs': 0.169792} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.541705] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.541971] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed3b2c5c-70fc-4c49-af8b-4e51dddb710b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.549802] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 972.549802] env[63293]: value = "task-1328020" [ 972.549802] env[63293]: _type = "Task" [ 972.549802] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.558377] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328020, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.622545] env[63293]: DEBUG nova.scheduler.client.report [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.870072] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.957623] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d456ca-3266-4380-9247-4d5a7c6dc200 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.983375] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53bbd301-a2ab-409a-8e5b-d83190742a27 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.007753] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance 'f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a' progress to 67 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 973.015019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "interface-35cbf277-b85a-4101-beae-0125fb5a61f5-c2fa0388-8ddf-497f-8200-903c0fde5a7b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.015019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-35cbf277-b85a-4101-beae-0125fb5a61f5-c2fa0388-8ddf-497f-8200-903c0fde5a7b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.015019] env[63293]: DEBUG nova.objects.instance [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'flavor' on Instance uuid 35cbf277-b85a-4101-beae-0125fb5a61f5 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.062180] env[63293]: DEBUG oslo_vmware.api [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328020, 'name': PowerOnVM_Task, 'duration_secs': 0.473949} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.062553] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.062760] env[63293]: INFO nova.compute.manager [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Took 10.42 seconds to spawn the instance on the hypervisor. [ 973.062942] env[63293]: DEBUG nova.compute.manager [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.063939] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0906a2b6-c225-487b-b976-31207224e04c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.071629] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.071797] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.071970] env[63293]: DEBUG nova.network.neutron [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 973.072177] env[63293]: DEBUG nova.objects.instance [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'info_cache' on Instance uuid 9ddf4f23-1279-4cbd-8212-10f344060445 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.127271] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.887s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.129548] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.696s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.129786] env[63293]: DEBUG nova.objects.instance [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lazy-loading 'resources' on Instance uuid 657d5f21-b985-4e3c-8799-b3aae2b1c509 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.149260] env[63293]: INFO nova.scheduler.client.report [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleted allocations for instance 66be1abd-c541-4b8a-8d72-98bde03b1888 [ 973.586554] env[63293]: DEBUG nova.network.neutron [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Port a20688c6-3fbe-4355-85f9-334019e37f43 binding to destination host cpu-1 is already ACTIVE {{(pid=63293) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 973.590805] env[63293]: INFO nova.compute.manager [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Took 22.10 seconds to build instance. [ 973.637712] env[63293]: DEBUG nova.objects.instance [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'pci_requests' on Instance uuid 35cbf277-b85a-4101-beae-0125fb5a61f5 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.656969] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6ba5f3c9-282e-473f-85f5-2adf82afdaf1 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "66be1abd-c541-4b8a-8d72-98bde03b1888" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.322s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.817699] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d9b776-9dc1-40f3-807d-51d065ce321d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.827031] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d098aa92-9049-4056-94eb-0c563e970a76 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.858186] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5199e256-ad01-4100-a767-126dd7018500 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.866597] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68630b31-6221-4d4e-a23e-8be9552631c8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.880474] env[63293]: DEBUG nova.compute.provider_tree [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.093284] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21e1b11b-d342-4400-8c35-faca05b95eb1 tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.613s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.142566] env[63293]: DEBUG nova.objects.base [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Object Instance<35cbf277-b85a-4101-beae-0125fb5a61f5> lazy-loaded attributes: flavor,pci_requests {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 974.142754] env[63293]: DEBUG nova.network.neutron [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 974.262483] env[63293]: DEBUG nova.policy [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2312d48038d7474481a0709267645780', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '866b347100454019b07f63922b995bb6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 974.384376] env[63293]: DEBUG nova.scheduler.client.report [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.404672] env[63293]: DEBUG nova.network.neutron [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance_info_cache with network_info: [{"id": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "address": "fa:16:3e:93:d3:25", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dbccf6a-05", "ovs_interfaceid": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.610409] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.610409] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.610772] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.683685] env[63293]: DEBUG nova.compute.manager [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.684750] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0b8b40-3ce0-4ca0-a586-e8e6e6bcb482 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.809699] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "d605e066-806a-4aac-9af6-f79d542ab2e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.809935] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.889356] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.760s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.892462] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.022s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.894476] env[63293]: INFO nova.compute.claims [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.907346] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.907623] env[63293]: DEBUG nova.objects.instance [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'migration_context' on Instance uuid 9ddf4f23-1279-4cbd-8212-10f344060445 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.913479] env[63293]: INFO nova.scheduler.client.report [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted allocations for instance 657d5f21-b985-4e3c-8799-b3aae2b1c509 [ 975.197060] env[63293]: INFO nova.compute.manager [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] instance snapshotting [ 975.199467] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2dcbcc1-9601-4913-9dad-1778a9bf9116 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.226682] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9effea-4bba-4b7d-9278-3f9b88460e0a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.314302] env[63293]: DEBUG nova.compute.manager [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 975.411855] env[63293]: DEBUG nova.objects.base [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Object Instance<9ddf4f23-1279-4cbd-8212-10f344060445> lazy-loaded attributes: info_cache,migration_context {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 975.413107] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06056eda-106e-4746-83d7-dfeba50fdb15 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.434463] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbd09058-7dce-4227-b729-259af5b7395b tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "657d5f21-b985-4e3c-8799-b3aae2b1c509" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.935s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.435373] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41323d3b-eb03-4bcf-b54f-d25c29ef005c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.442130] env[63293]: DEBUG oslo_vmware.api [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 975.442130] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52198d93-6e30-e0c6-bbe8-465dc0d419c9" [ 975.442130] env[63293]: _type = "Task" [ 975.442130] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.453501] env[63293]: DEBUG oslo_vmware.api [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52198d93-6e30-e0c6-bbe8-465dc0d419c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.601864] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "828180c0-b7f8-4666-9f5d-e6741ef6495a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.602988] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "828180c0-b7f8-4666-9f5d-e6741ef6495a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.602988] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "828180c0-b7f8-4666-9f5d-e6741ef6495a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.602988] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "828180c0-b7f8-4666-9f5d-e6741ef6495a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.602988] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "828180c0-b7f8-4666-9f5d-e6741ef6495a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.606521] env[63293]: INFO nova.compute.manager [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Terminating instance [ 975.607321] env[63293]: DEBUG nova.compute.manager [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 975.607641] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 975.608606] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1cdeff5-bac8-45e1-b025-b63104a917f5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.620062] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 975.620166] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-546f341a-a790-4d1b-8d1d-1f831168db78 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.637183] env[63293]: DEBUG oslo_vmware.api [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 975.637183] env[63293]: value = "task-1328021" [ 975.637183] env[63293]: _type = "Task" [ 975.637183] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.646776] env[63293]: DEBUG oslo_vmware.api [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328021, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.659711] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.659885] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.660170] env[63293]: DEBUG nova.network.neutron [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 975.742837] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Creating Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 975.744050] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c7035ee1-e1ec-41d0-adff-815318c5e138 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.755940] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 975.755940] env[63293]: value = "task-1328022" [ 975.755940] env[63293]: _type = "Task" [ 975.755940] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.763178] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328022, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.840769] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.892029] env[63293]: DEBUG nova.compute.manager [req-c3566eb0-ac0e-4f19-8a88-87556396a5d2 req-1c618689-ba15-41e2-8f4a-761e2d68fe65 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Received event network-vif-plugged-c2fa0388-8ddf-497f-8200-903c0fde5a7b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.892307] env[63293]: DEBUG oslo_concurrency.lockutils [req-c3566eb0-ac0e-4f19-8a88-87556396a5d2 req-1c618689-ba15-41e2-8f4a-761e2d68fe65 service nova] Acquiring lock "35cbf277-b85a-4101-beae-0125fb5a61f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.892816] env[63293]: DEBUG oslo_concurrency.lockutils [req-c3566eb0-ac0e-4f19-8a88-87556396a5d2 req-1c618689-ba15-41e2-8f4a-761e2d68fe65 service nova] Lock "35cbf277-b85a-4101-beae-0125fb5a61f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.893015] env[63293]: DEBUG oslo_concurrency.lockutils [req-c3566eb0-ac0e-4f19-8a88-87556396a5d2 req-1c618689-ba15-41e2-8f4a-761e2d68fe65 service nova] Lock "35cbf277-b85a-4101-beae-0125fb5a61f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.893210] env[63293]: DEBUG nova.compute.manager [req-c3566eb0-ac0e-4f19-8a88-87556396a5d2 req-1c618689-ba15-41e2-8f4a-761e2d68fe65 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] No waiting events found dispatching network-vif-plugged-c2fa0388-8ddf-497f-8200-903c0fde5a7b {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 975.893383] env[63293]: WARNING nova.compute.manager [req-c3566eb0-ac0e-4f19-8a88-87556396a5d2 req-1c618689-ba15-41e2-8f4a-761e2d68fe65 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Received unexpected event network-vif-plugged-c2fa0388-8ddf-497f-8200-903c0fde5a7b for instance with vm_state active and task_state None. [ 975.914188] env[63293]: DEBUG nova.network.neutron [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Successfully updated port: c2fa0388-8ddf-497f-8200-903c0fde5a7b {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.953098] env[63293]: DEBUG oslo_vmware.api [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52198d93-6e30-e0c6-bbe8-465dc0d419c9, 'name': SearchDatastore_Task, 'duration_secs': 0.013746} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.955761] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.083746] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e35da6-cf32-4f9b-b24b-26d7e12f6a53 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.092948] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01c7bad-b974-4b56-a364-ed69051975fe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.124345] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ae8426-863f-42a5-b98d-b7e742184f18 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.132437] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc782ee-2c50-4c25-b138-308c5cdc85c7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.840262] env[63293]: DEBUG oslo_concurrency.lockutils [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.840451] env[63293]: DEBUG oslo_concurrency.lockutils [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.840616] env[63293]: DEBUG nova.network.neutron [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 976.862472] env[63293]: DEBUG nova.compute.provider_tree [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.865178] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328022, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.867120] env[63293]: DEBUG oslo_vmware.api [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328021, 'name': PowerOffVM_Task, 'duration_secs': 0.222081} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.867590] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 976.867770] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 976.868076] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f2473fa-1681-4ea1-b7a8-21628c43b83d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.052199] env[63293]: DEBUG nova.network.neutron [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance_info_cache with network_info: [{"id": "a20688c6-3fbe-4355-85f9-334019e37f43", "address": "fa:16:3e:08:30:f7", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa20688c6-3f", "ovs_interfaceid": "a20688c6-3fbe-4355-85f9-334019e37f43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.351516] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328022, 'name': CreateSnapshot_Task, 'duration_secs': 1.245404} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.351800] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Created Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 977.352831] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371aab06-1270-4f62-8866-1e6db95504b0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.368364] env[63293]: DEBUG nova.scheduler.client.report [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.387316] env[63293]: WARNING nova.network.neutron [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] 908b99b3-6031-4de0-86a4-09a9cbe355a5 already exists in list: networks containing: ['908b99b3-6031-4de0-86a4-09a9cbe355a5']. ignoring it [ 977.555159] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.664405] env[63293]: DEBUG nova.network.neutron [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updating instance_info_cache with network_info: [{"id": "df1c83c6-23d7-439b-a496-234c48738d81", "address": "fa:16:3e:0f:9f:14", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1c83c6-23", "ovs_interfaceid": "df1c83c6-23d7-439b-a496-234c48738d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2fa0388-8ddf-497f-8200-903c0fde5a7b", "address": "fa:16:3e:ae:52:3b", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2fa0388-8d", "ovs_interfaceid": "c2fa0388-8ddf-497f-8200-903c0fde5a7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.870413] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Creating linked-clone VM from snapshot {{(pid=63293) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 977.870581] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4adc5e79-4a12-42f1-a163-cf04abe88b11 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.874199] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.982s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.874658] env[63293]: DEBUG nova.compute.manager [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 977.877353] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.037s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.878771] env[63293]: INFO nova.compute.claims [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.887824] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 977.887824] env[63293]: value = "task-1328024" [ 977.887824] env[63293]: _type = "Task" [ 977.887824] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.897168] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328024, 'name': CloneVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.025773] env[63293]: DEBUG nova.compute.manager [req-723dbf5f-8e54-45a9-a0f5-0306299aa749 req-f192dbd4-ea0a-48de-9e21-299c8ad363a2 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Received event network-changed-c2fa0388-8ddf-497f-8200-903c0fde5a7b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.025939] env[63293]: DEBUG nova.compute.manager [req-723dbf5f-8e54-45a9-a0f5-0306299aa749 req-f192dbd4-ea0a-48de-9e21-299c8ad363a2 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing instance network info cache due to event network-changed-c2fa0388-8ddf-497f-8200-903c0fde5a7b. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 978.026150] env[63293]: DEBUG oslo_concurrency.lockutils [req-723dbf5f-8e54-45a9-a0f5-0306299aa749 req-f192dbd4-ea0a-48de-9e21-299c8ad363a2 service nova] Acquiring lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.077394] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cceda56-0977-4a5f-b413-dce727e1bb8e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.103816] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e1066b-3ada-4f6d-a286-3d263fabe915 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.111514] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance 'f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a' progress to 83 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 978.169215] env[63293]: DEBUG oslo_concurrency.lockutils [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.169868] env[63293]: DEBUG oslo_concurrency.lockutils [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.170278] env[63293]: DEBUG oslo_concurrency.lockutils [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.171669] env[63293]: DEBUG oslo_concurrency.lockutils [req-723dbf5f-8e54-45a9-a0f5-0306299aa749 req-f192dbd4-ea0a-48de-9e21-299c8ad363a2 service nova] Acquired lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.171669] env[63293]: DEBUG nova.network.neutron [req-723dbf5f-8e54-45a9-a0f5-0306299aa749 req-f192dbd4-ea0a-48de-9e21-299c8ad363a2 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing network info cache for port c2fa0388-8ddf-497f-8200-903c0fde5a7b {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 978.173208] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b95ceb-f09e-42d1-af42-34754b70aa96 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.195627] env[63293]: DEBUG nova.virt.hardware [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 978.195880] env[63293]: DEBUG nova.virt.hardware [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 978.196056] env[63293]: DEBUG nova.virt.hardware [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 978.196246] env[63293]: DEBUG nova.virt.hardware [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 978.196406] env[63293]: DEBUG nova.virt.hardware [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 978.196555] env[63293]: DEBUG nova.virt.hardware [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 978.196764] env[63293]: DEBUG nova.virt.hardware [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 978.196928] env[63293]: DEBUG nova.virt.hardware [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 978.197171] env[63293]: DEBUG nova.virt.hardware [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 978.197374] env[63293]: DEBUG nova.virt.hardware [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 978.197554] env[63293]: DEBUG nova.virt.hardware [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 978.206950] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Reconfiguring VM to attach interface {{(pid=63293) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 978.207627] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b13e549b-9ff0-4c8f-ab27-50d49c4e037e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.227371] env[63293]: DEBUG oslo_vmware.api [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 978.227371] env[63293]: value = "task-1328025" [ 978.227371] env[63293]: _type = "Task" [ 978.227371] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.237203] env[63293]: DEBUG oslo_vmware.api [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328025, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.386736] env[63293]: DEBUG nova.compute.utils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 978.388325] env[63293]: DEBUG nova.compute.manager [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 978.388463] env[63293]: DEBUG nova.network.neutron [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 978.400282] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328024, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.435509] env[63293]: DEBUG nova.policy [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5dab7cefbc2c4360b927f4c7e5e5319e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e1a30d37c1b47fd8670db0a9ef7020d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 978.618923] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.619532] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5c4c609-8c7c-4196-9530-abc5880d8ab8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.628551] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 978.628551] env[63293]: value = "task-1328026" [ 978.628551] env[63293]: _type = "Task" [ 978.628551] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.639891] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328026, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.740421] env[63293]: DEBUG oslo_vmware.api [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328025, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.762197] env[63293]: DEBUG nova.network.neutron [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Successfully created port: cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.893566] env[63293]: DEBUG nova.compute.manager [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 978.906977] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328024, 'name': CloneVM_Task} progress is 95%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.045547] env[63293]: DEBUG nova.network.neutron [req-723dbf5f-8e54-45a9-a0f5-0306299aa749 req-f192dbd4-ea0a-48de-9e21-299c8ad363a2 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updated VIF entry in instance network info cache for port c2fa0388-8ddf-497f-8200-903c0fde5a7b. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 979.045972] env[63293]: DEBUG nova.network.neutron [req-723dbf5f-8e54-45a9-a0f5-0306299aa749 req-f192dbd4-ea0a-48de-9e21-299c8ad363a2 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updating instance_info_cache with network_info: [{"id": "df1c83c6-23d7-439b-a496-234c48738d81", "address": "fa:16:3e:0f:9f:14", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1c83c6-23", "ovs_interfaceid": "df1c83c6-23d7-439b-a496-234c48738d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2fa0388-8ddf-497f-8200-903c0fde5a7b", "address": "fa:16:3e:ae:52:3b", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2fa0388-8d", "ovs_interfaceid": "c2fa0388-8ddf-497f-8200-903c0fde5a7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.114403] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 979.114746] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 979.114977] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleting the datastore file [datastore1] 828180c0-b7f8-4666-9f5d-e6741ef6495a {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.115595] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f93f5e1a-04c9-4f16-b00f-22e2c82c573e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.127300] env[63293]: DEBUG oslo_vmware.api [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 979.127300] env[63293]: value = "task-1328027" [ 979.127300] env[63293]: _type = "Task" [ 979.127300] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.127576] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104ba695-8718-4d90-b242-88bf7b9668cd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.144447] env[63293]: DEBUG oslo_vmware.api [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328027, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.150075] env[63293]: DEBUG oslo_vmware.api [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328026, 'name': PowerOnVM_Task, 'duration_secs': 0.501332} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.151349] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9c590a-6ca8-4187-8262-46901307f9e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.155266] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 979.155518] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9947a5a9-bafc-4df1-adaf-445b7aad7495 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance 'f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a' progress to 100 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 979.190664] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce60547f-e7c6-453e-9fa7-60e129a05012 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.201024] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ace3e4-d1a7-4eae-9e39-a10a1a4efe6d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.213676] env[63293]: DEBUG nova.compute.provider_tree [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 979.240341] env[63293]: DEBUG oslo_vmware.api [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328025, 'name': ReconfigVM_Task, 'duration_secs': 0.964237} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.240887] env[63293]: DEBUG oslo_concurrency.lockutils [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.241164] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Reconfigured VM to attach interface {{(pid=63293) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 979.406305] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328024, 'name': CloneVM_Task, 'duration_secs': 1.391822} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.406761] env[63293]: INFO nova.virt.vmwareapi.vmops [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Created linked-clone VM from snapshot [ 979.407506] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1592b3-3ee8-4a9f-b85d-c7c7df91a587 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.415848] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Uploading image 7b388520-9bc0-45ff-ba2d-fb3eb9074f95 {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 979.429492] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Destroying the VM {{(pid=63293) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 979.429818] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-60df54f9-fc6f-43d1-b19f-478e61dc8222 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.439820] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 979.439820] env[63293]: value = "task-1328028" [ 979.439820] env[63293]: _type = "Task" [ 979.439820] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.449216] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328028, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.549955] env[63293]: DEBUG oslo_concurrency.lockutils [req-723dbf5f-8e54-45a9-a0f5-0306299aa749 req-f192dbd4-ea0a-48de-9e21-299c8ad363a2 service nova] Releasing lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.639236] env[63293]: DEBUG oslo_vmware.api [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328027, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.474852} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.639707] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.640040] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 979.640295] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 979.640559] env[63293]: INFO nova.compute.manager [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Took 4.03 seconds to destroy the instance on the hypervisor. [ 979.640904] env[63293]: DEBUG oslo.service.loopingcall [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.641193] env[63293]: DEBUG nova.compute.manager [-] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 979.641331] env[63293]: DEBUG nova.network.neutron [-] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 979.734246] env[63293]: ERROR nova.scheduler.client.report [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [req-eb27387e-d48a-4dd3-bd48-54f8a8fb8215] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 619725c3-6194-4724-94b1-1dd499be55d0. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-eb27387e-d48a-4dd3-bd48-54f8a8fb8215"}]} [ 979.745540] env[63293]: DEBUG oslo_concurrency.lockutils [None req-460a492c-abf2-4667-a576-6657e6551b3e tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-35cbf277-b85a-4101-beae-0125fb5a61f5-c2fa0388-8ddf-497f-8200-903c0fde5a7b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.734s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.751942] env[63293]: DEBUG nova.scheduler.client.report [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Refreshing inventories for resource provider 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 979.766976] env[63293]: DEBUG nova.scheduler.client.report [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Updating ProviderTree inventory for provider 619725c3-6194-4724-94b1-1dd499be55d0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 979.767224] env[63293]: DEBUG nova.compute.provider_tree [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Updating inventory in ProviderTree for provider 619725c3-6194-4724-94b1-1dd499be55d0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 979.779187] env[63293]: DEBUG nova.scheduler.client.report [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Refreshing aggregate associations for resource provider 619725c3-6194-4724-94b1-1dd499be55d0, aggregates: None {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 979.799220] env[63293]: DEBUG nova.scheduler.client.report [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Refreshing trait associations for resource provider 619725c3-6194-4724-94b1-1dd499be55d0, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=63293) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 979.910014] env[63293]: DEBUG nova.compute.manager [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 979.937424] env[63293]: DEBUG nova.virt.hardware [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.937699] env[63293]: DEBUG nova.virt.hardware [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.937863] env[63293]: DEBUG nova.virt.hardware [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.938060] env[63293]: DEBUG nova.virt.hardware [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.938217] env[63293]: DEBUG nova.virt.hardware [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.938365] env[63293]: DEBUG nova.virt.hardware [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.938712] env[63293]: DEBUG nova.virt.hardware [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.938889] env[63293]: DEBUG nova.virt.hardware [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.939075] env[63293]: DEBUG nova.virt.hardware [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.939249] env[63293]: DEBUG nova.virt.hardware [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.939480] env[63293]: DEBUG nova.virt.hardware [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.940455] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7481877-9b85-455b-b608-a58365f8cd6a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.954617] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328028, 'name': Destroy_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.957780] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8526ebd-983e-4a02-ae87-6ae6449253a2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.000136] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd29b5f4-fd7f-426f-b58e-9932fa0e8366 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.008327] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07480d9-e1b4-4f20-8284-7a405785ed67 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.039925] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d74a2b-6973-4106-941e-94ba9b84a03a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.048340] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eddd04b-1a5d-4b28-a086-1bc99e3e4ca6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.064431] env[63293]: DEBUG nova.compute.provider_tree [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.192568] env[63293]: DEBUG nova.compute.manager [req-f4b8ea4a-4342-4a16-9613-13ffa3915027 req-82a78ccd-e300-4ec5-8b90-12e7972814a1 service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Received event network-vif-deleted-c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.192826] env[63293]: INFO nova.compute.manager [req-f4b8ea4a-4342-4a16-9613-13ffa3915027 req-82a78ccd-e300-4ec5-8b90-12e7972814a1 service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Neutron deleted interface c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242; detaching it from the instance and deleting it from the info cache [ 980.192970] env[63293]: DEBUG nova.network.neutron [req-f4b8ea4a-4342-4a16-9613-13ffa3915027 req-82a78ccd-e300-4ec5-8b90-12e7972814a1 service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.393048] env[63293]: DEBUG nova.network.neutron [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Successfully updated port: cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 980.403114] env[63293]: DEBUG nova.network.neutron [-] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.450997] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328028, 'name': Destroy_Task, 'duration_secs': 0.827215} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.451388] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Destroyed the VM [ 980.451672] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Deleting Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 980.451933] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-330921d1-fcc2-4cbb-8a82-00b2fd053334 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.459733] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 980.459733] env[63293]: value = "task-1328029" [ 980.459733] env[63293]: _type = "Task" [ 980.459733] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.467908] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328029, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.571122] env[63293]: DEBUG nova.scheduler.client.report [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.696120] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f62cf347-46ff-4ade-8b63-bfe8973985de {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.708794] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df950f5-b780-44b6-93d9-607d1b5207b9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.741365] env[63293]: DEBUG nova.compute.manager [req-f4b8ea4a-4342-4a16-9613-13ffa3915027 req-82a78ccd-e300-4ec5-8b90-12e7972814a1 service nova] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Detach interface failed, port_id=c1fa4ad6-2dd0-49ee-8625-9ba8cff4e242, reason: Instance 828180c0-b7f8-4666-9f5d-e6741ef6495a could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 980.896373] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "refresh_cache-9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.896565] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquired lock "refresh_cache-9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.896737] env[63293]: DEBUG nova.network.neutron [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 980.904796] env[63293]: INFO nova.compute.manager [-] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Took 1.26 seconds to deallocate network for instance. [ 980.969789] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328029, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.076381] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.199s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.076944] env[63293]: DEBUG nova.compute.manager [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 981.079680] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 5.124s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.271634] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "interface-35cbf277-b85a-4101-beae-0125fb5a61f5-c2fa0388-8ddf-497f-8200-903c0fde5a7b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.272055] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-35cbf277-b85a-4101-beae-0125fb5a61f5-c2fa0388-8ddf-497f-8200-903c0fde5a7b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.411952] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.430942] env[63293]: DEBUG nova.network.neutron [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 981.471953] env[63293]: DEBUG oslo_vmware.api [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328029, 'name': RemoveSnapshot_Task, 'duration_secs': 0.539589} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.472266] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Deleted Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 981.588918] env[63293]: DEBUG nova.compute.utils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.591932] env[63293]: DEBUG nova.compute.manager [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 981.592152] env[63293]: DEBUG nova.network.neutron [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 981.683293] env[63293]: DEBUG nova.network.neutron [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Updating instance_info_cache with network_info: [{"id": "cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3", "address": "fa:16:3e:24:8c:84", "network": {"id": "9d5a3e58-567a-4c3f-9ad9-9fcbf81fced9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-949261114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1a30d37c1b47fd8670db0a9ef7020d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc516442-7d", "ovs_interfaceid": "cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.696103] env[63293]: DEBUG nova.policy [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd2a9495ab0b4768a4075a71f25aa9a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff8479fc251544e9b4618deccf992754', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 981.774713] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.775014] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.776170] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe1f8cd-f200-44f5-9b0a-d63fec0877e7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.796798] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36072411-47a3-4c5b-8499-a174a323a163 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.824432] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Reconfiguring VM to detach interface {{(pid=63293) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 981.826984] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be039d24-164c-431e-a91b-67b1d8e1b3e7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.845020] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d11b34-56ce-4f99-8745-7f0b7f752f8e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.847152] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.847433] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.847619] env[63293]: DEBUG nova.compute.manager [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Going to confirm migration 3 {{(pid=63293) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 981.857021] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-408cf6a7-1afe-4724-bd79-94978a706bf2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.863051] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 981.863051] env[63293]: value = "task-1328030" [ 981.863051] env[63293]: _type = "Task" [ 981.863051] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.900346] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759ea82a-993f-4708-b25b-bd457c6fc11d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.903254] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.909240] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d59979-ac91-4bd3-ae4a-47ffe1f93879 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.924178] env[63293]: DEBUG nova.compute.provider_tree [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.979387] env[63293]: WARNING nova.compute.manager [None req-b8fe9176-543e-470a-afac-7bc8e431591b tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Image not found during snapshot: nova.exception.ImageNotFound: Image 7b388520-9bc0-45ff-ba2d-fb3eb9074f95 could not be found. [ 982.018582] env[63293]: DEBUG nova.network.neutron [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Successfully created port: 83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 982.095641] env[63293]: DEBUG nova.compute.manager [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 982.186073] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Releasing lock "refresh_cache-9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.186419] env[63293]: DEBUG nova.compute.manager [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Instance network_info: |[{"id": "cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3", "address": "fa:16:3e:24:8c:84", "network": {"id": "9d5a3e58-567a-4c3f-9ad9-9fcbf81fced9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-949261114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1a30d37c1b47fd8670db0a9ef7020d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc516442-7d", "ovs_interfaceid": "cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 982.186830] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:8c:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.194503] env[63293]: DEBUG oslo.service.loopingcall [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.194950] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 982.195289] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b2b4e7e-9f24-42a2-9942-79a8b089e57e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.219051] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.219051] env[63293]: value = "task-1328031" [ 982.219051] env[63293]: _type = "Task" [ 982.219051] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.221342] env[63293]: DEBUG nova.compute.manager [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Received event network-vif-plugged-cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.221553] env[63293]: DEBUG oslo_concurrency.lockutils [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] Acquiring lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.221761] env[63293]: DEBUG oslo_concurrency.lockutils [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.222241] env[63293]: DEBUG oslo_concurrency.lockutils [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.222241] env[63293]: DEBUG nova.compute.manager [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] No waiting events found dispatching network-vif-plugged-cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 982.222352] env[63293]: WARNING nova.compute.manager [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Received unexpected event network-vif-plugged-cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3 for instance with vm_state building and task_state spawning. [ 982.222494] env[63293]: DEBUG nova.compute.manager [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Received event network-changed-cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.222660] env[63293]: DEBUG nova.compute.manager [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Refreshing instance network info cache due to event network-changed-cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.222843] env[63293]: DEBUG oslo_concurrency.lockutils [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] Acquiring lock "refresh_cache-9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.222983] env[63293]: DEBUG oslo_concurrency.lockutils [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] Acquired lock "refresh_cache-9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.223156] env[63293]: DEBUG nova.network.neutron [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Refreshing network info cache for port cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 982.233630] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328031, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.344465] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.344885] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.344965] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.345141] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.345317] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.347559] env[63293]: INFO nova.compute.manager [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Terminating instance [ 982.349427] env[63293]: DEBUG nova.compute.manager [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 982.349625] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 982.350643] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d792a082-aca5-43d1-8449-51ed49eb13d3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.361608] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.361845] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d03bebbc-9e00-483e-a616-877701c81ef5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.371360] env[63293]: DEBUG oslo_vmware.api [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 982.371360] env[63293]: value = "task-1328032" [ 982.371360] env[63293]: _type = "Task" [ 982.371360] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.377901] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.383326] env[63293]: DEBUG oslo_vmware.api [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328032, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.417106] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.417359] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.417565] env[63293]: DEBUG nova.network.neutron [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 982.417772] env[63293]: DEBUG nova.objects.instance [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lazy-loading 'info_cache' on Instance uuid f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.428123] env[63293]: DEBUG nova.scheduler.client.report [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.660233] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "f429869f-9e9f-443e-a980-67a80c3bd799" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.660507] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "f429869f-9e9f-443e-a980-67a80c3bd799" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.660700] env[63293]: INFO nova.compute.manager [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Shelving [ 982.732812] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328031, 'name': CreateVM_Task, 'duration_secs': 0.428716} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.732988] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 982.733908] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.734086] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.734442] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.734988] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dae7c54-7daf-43f8-87c7-1d78fd1b7aaa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.740164] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 982.740164] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5270b28a-ab97-a915-df31-6d551fd7be97" [ 982.740164] env[63293]: _type = "Task" [ 982.740164] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.749129] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5270b28a-ab97-a915-df31-6d551fd7be97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.878481] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.883694] env[63293]: DEBUG oslo_vmware.api [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328032, 'name': PowerOffVM_Task, 'duration_secs': 0.242505} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.883963] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 982.884152] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 982.884401] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97e51f5b-7d19-459c-81c1-97777b79942c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.947244] env[63293]: DEBUG nova.network.neutron [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Updated VIF entry in instance network info cache for port cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 982.947786] env[63293]: DEBUG nova.network.neutron [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Updating instance_info_cache with network_info: [{"id": "cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3", "address": "fa:16:3e:24:8c:84", "network": {"id": "9d5a3e58-567a-4c3f-9ad9-9fcbf81fced9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-949261114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1a30d37c1b47fd8670db0a9ef7020d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc516442-7d", "ovs_interfaceid": "cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.957265] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 982.957833] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 982.958603] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleting the datastore file [datastore1] 9d3fcf27-2c2c-4975-a802-a5c0859640d0 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 982.959087] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-642b8c70-460d-4fe2-917a-518c1093b5d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.966364] env[63293]: DEBUG oslo_vmware.api [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for the task: (returnval){ [ 982.966364] env[63293]: value = "task-1328034" [ 982.966364] env[63293]: _type = "Task" [ 982.966364] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.975272] env[63293]: DEBUG oslo_vmware.api [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328034, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.105694] env[63293]: DEBUG nova.compute.manager [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 983.131703] env[63293]: DEBUG nova.virt.hardware [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.132016] env[63293]: DEBUG nova.virt.hardware [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.132216] env[63293]: DEBUG nova.virt.hardware [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.132452] env[63293]: DEBUG nova.virt.hardware [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.132645] env[63293]: DEBUG nova.virt.hardware [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.132836] env[63293]: DEBUG nova.virt.hardware [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.133118] env[63293]: DEBUG nova.virt.hardware [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.133328] env[63293]: DEBUG nova.virt.hardware [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.133538] env[63293]: DEBUG nova.virt.hardware [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.133749] env[63293]: DEBUG nova.virt.hardware [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.133966] env[63293]: DEBUG nova.virt.hardware [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.134847] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf1c4fca-83a4-46ff-95f3-26aba8dce901 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.143298] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288bd165-ecbb-4483-82f2-660a2ba511d8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.167838] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 983.168189] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31e29285-7e11-42a3-82b0-90bd5714cbaa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.178054] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 983.178054] env[63293]: value = "task-1328035" [ 983.178054] env[63293]: _type = "Task" [ 983.178054] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.188474] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328035, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.252499] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5270b28a-ab97-a915-df31-6d551fd7be97, 'name': SearchDatastore_Task, 'duration_secs': 0.024149} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.252817] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.253068] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.253311] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.253459] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.253641] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.253914] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5328b0d6-bfd6-4b71-9c83-96e5846e0337 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.262963] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.263170] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 983.263936] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f23ecdf6-c49d-43d3-b773-5ee857a1aee9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.269643] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 983.269643] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52db2e41-f8ad-c51b-ec24-df0226fa055f" [ 983.269643] env[63293]: _type = "Task" [ 983.269643] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.277830] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52db2e41-f8ad-c51b-ec24-df0226fa055f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.376637] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.438862] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.359s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.439133] env[63293]: DEBUG nova.compute.manager [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=63293) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 983.442077] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.030s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.442318] env[63293]: DEBUG nova.objects.instance [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lazy-loading 'resources' on Instance uuid 828180c0-b7f8-4666-9f5d-e6741ef6495a {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.450637] env[63293]: DEBUG oslo_concurrency.lockutils [req-d640df97-dce0-4088-bdaa-44e7ac95c378 req-85f07ca8-72af-4f31-9f56-ae86e0b18f9a service nova] Releasing lock "refresh_cache-9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.477210] env[63293]: DEBUG oslo_vmware.api [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Task: {'id': task-1328034, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.289408} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.479645] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 983.479850] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 983.480089] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 983.480301] env[63293]: INFO nova.compute.manager [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Took 1.13 seconds to destroy the instance on the hypervisor. [ 983.480550] env[63293]: DEBUG oslo.service.loopingcall [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.481487] env[63293]: DEBUG nova.compute.manager [-] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 983.481611] env[63293]: DEBUG nova.network.neutron [-] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 983.630722] env[63293]: DEBUG nova.network.neutron [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Successfully updated port: 83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.692341] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328035, 'name': PowerOffVM_Task, 'duration_secs': 0.241241} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.692341] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.692341] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadb9cf3-b139-43c7-a4d3-38ee28766f54 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.716838] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9be852b-6215-4ee1-ab51-78f664d65802 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.719567] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "ad585ebb-2072-45df-b645-94c9fa93576b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.720069] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.781550] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52db2e41-f8ad-c51b-ec24-df0226fa055f, 'name': SearchDatastore_Task, 'duration_secs': 0.010675} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.782594] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1916638b-3505-4924-ad71-4289ca0752c6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.788656] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 983.788656] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5251b8f9-7229-c2bf-93d9-be112ac1a193" [ 983.788656] env[63293]: _type = "Task" [ 983.788656] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.789645] env[63293]: DEBUG nova.network.neutron [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance_info_cache with network_info: [{"id": "a20688c6-3fbe-4355-85f9-334019e37f43", "address": "fa:16:3e:08:30:f7", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa20688c6-3f", "ovs_interfaceid": "a20688c6-3fbe-4355-85f9-334019e37f43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.800389] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5251b8f9-7229-c2bf-93d9-be112ac1a193, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.877787] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.920372] env[63293]: DEBUG nova.compute.manager [req-56c564c1-3309-426d-8ea7-80742eb912c1 req-e3230775-cc40-4f0d-abbc-a0e038d27e2a service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Received event network-vif-deleted-bda9acec-71cb-4de2-859e-75837898a336 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.920530] env[63293]: INFO nova.compute.manager [req-56c564c1-3309-426d-8ea7-80742eb912c1 req-e3230775-cc40-4f0d-abbc-a0e038d27e2a service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Neutron deleted interface bda9acec-71cb-4de2-859e-75837898a336; detaching it from the instance and deleting it from the info cache [ 983.920708] env[63293]: DEBUG nova.network.neutron [req-56c564c1-3309-426d-8ea7-80742eb912c1 req-e3230775-cc40-4f0d-abbc-a0e038d27e2a service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.004058] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.004341] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.004565] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.004736] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.004906] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.007223] env[63293]: INFO nova.scheduler.client.report [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleted allocation for migration a66e3649-f4ed-4fb1-8bca-bb060f430e05 [ 984.008658] env[63293]: INFO nova.compute.manager [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Terminating instance [ 984.012018] env[63293]: DEBUG nova.compute.manager [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.012235] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.013283] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26fb1272-3c1c-4d1e-b0af-0b38cf3ab197 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.024643] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.025019] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5077b0ad-8894-422b-9970-bfafb6088f46 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.034025] env[63293]: DEBUG oslo_vmware.api [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 984.034025] env[63293]: value = "task-1328036" [ 984.034025] env[63293]: _type = "Task" [ 984.034025] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.043619] env[63293]: DEBUG oslo_vmware.api [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328036, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.134288] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.134482] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.134894] env[63293]: DEBUG nova.network.neutron [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 984.148741] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bebbaa0-ed5c-48a2-9870-c60acd14392c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.157542] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e1853f-e685-4cc7-917f-893bd9c46742 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.191571] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101456d4-e5e8-4cc3-a741-cf18e8c369b1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.200761] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e081a3df-53ce-48b0-b01f-9ea662b2c75b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.216491] env[63293]: DEBUG nova.compute.provider_tree [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.222619] env[63293]: INFO nova.compute.manager [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Detaching volume 658fbd49-f54e-404a-99c2-e7147e4a3700 [ 984.234872] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Creating Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 984.235481] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8cc4aa66-3dad-40bc-854c-9ab31b4ed8e6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.249026] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 984.249026] env[63293]: value = "task-1328037" [ 984.249026] env[63293]: _type = "Task" [ 984.249026] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.252142] env[63293]: DEBUG nova.compute.manager [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Received event network-vif-plugged-83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.252384] env[63293]: DEBUG oslo_concurrency.lockutils [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] Acquiring lock "d605e066-806a-4aac-9af6-f79d542ab2e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.252695] env[63293]: DEBUG oslo_concurrency.lockutils [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.252802] env[63293]: DEBUG oslo_concurrency.lockutils [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.252982] env[63293]: DEBUG nova.compute.manager [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] No waiting events found dispatching network-vif-plugged-83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 984.253186] env[63293]: WARNING nova.compute.manager [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Received unexpected event network-vif-plugged-83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970 for instance with vm_state building and task_state spawning. [ 984.253341] env[63293]: DEBUG nova.compute.manager [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Received event network-changed-83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.253549] env[63293]: DEBUG nova.compute.manager [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Refreshing instance network info cache due to event network-changed-83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 984.253743] env[63293]: DEBUG oslo_concurrency.lockutils [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] Acquiring lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.261074] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328037, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.265490] env[63293]: INFO nova.virt.block_device [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Attempting to driver detach volume 658fbd49-f54e-404a-99c2-e7147e4a3700 from mountpoint /dev/sdb [ 984.265747] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Volume detach. Driver type: vmdk {{(pid=63293) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 984.265959] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283793', 'volume_id': '658fbd49-f54e-404a-99c2-e7147e4a3700', 'name': 'volume-658fbd49-f54e-404a-99c2-e7147e4a3700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'ad585ebb-2072-45df-b645-94c9fa93576b', 'attached_at': '', 'detached_at': '', 'volume_id': '658fbd49-f54e-404a-99c2-e7147e4a3700', 'serial': '658fbd49-f54e-404a-99c2-e7147e4a3700'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 984.266983] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce56a00c-7c0c-4a19-b586-02c021a1871a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.294143] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68134895-66bf-4581-8da5-a4f1ee0df0d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.297315] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.297576] env[63293]: DEBUG nova.objects.instance [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lazy-loading 'migration_context' on Instance uuid f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.306569] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5251b8f9-7229-c2bf-93d9-be112ac1a193, 'name': SearchDatastore_Task, 'duration_secs': 0.01548} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.309216] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.309507] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2/9937a3c4-c3a6-423b-ac4f-50b7ec5879d2.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 984.310219] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f1cda42-9053-435e-ab9d-04d84f8e0c97 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.312763] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf345b47-f163-4b93-819a-dfef9d951e0f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.339970] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3864b6-3d72-4bb2-b630-66fc0cbb01ff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.342892] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 984.342892] env[63293]: value = "task-1328038" [ 984.342892] env[63293]: _type = "Task" [ 984.342892] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.357610] env[63293]: DEBUG nova.network.neutron [-] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.359554] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] The volume has not been displaced from its original location: [datastore1] volume-658fbd49-f54e-404a-99c2-e7147e4a3700/volume-658fbd49-f54e-404a-99c2-e7147e4a3700.vmdk. No consolidation needed. {{(pid=63293) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 984.365066] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Reconfiguring VM instance instance-0000003f to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 984.366127] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6307a596-5f2a-4272-98b6-b4b992859126 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.385458] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328038, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.395405] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.396790] env[63293]: DEBUG oslo_vmware.api [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 984.396790] env[63293]: value = "task-1328039" [ 984.396790] env[63293]: _type = "Task" [ 984.396790] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.405946] env[63293]: DEBUG oslo_vmware.api [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328039, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.423401] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5b69506b-8aca-496c-9621-f6973411cb4d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.434188] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0c8644-49f1-4ff6-9fa2-768f4e04f9c9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.467784] env[63293]: DEBUG nova.compute.manager [req-56c564c1-3309-426d-8ea7-80742eb912c1 req-e3230775-cc40-4f0d-abbc-a0e038d27e2a service nova] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Detach interface failed, port_id=bda9acec-71cb-4de2-859e-75837898a336, reason: Instance 9d3fcf27-2c2c-4975-a802-a5c0859640d0 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 984.496937] env[63293]: DEBUG nova.objects.instance [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'flavor' on Instance uuid 9ddf4f23-1279-4cbd-8212-10f344060445 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.520189] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9b32efc0-fa91-4f96-8138-3b260fc28ce7 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 12.217s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.545273] env[63293]: DEBUG oslo_vmware.api [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328036, 'name': PowerOffVM_Task, 'duration_secs': 0.274536} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.545597] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.545777] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 984.546060] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e00c2cbc-81aa-4ae0-b5bb-6c20978eff46 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.669762] env[63293]: DEBUG nova.network.neutron [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 984.703696] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 984.703861] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 984.704319] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleting the datastore file [datastore1] 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.704465] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca5a9cfb-c839-46fb-a2fa-76e03dba7a7b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.715190] env[63293]: DEBUG oslo_vmware.api [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 984.715190] env[63293]: value = "task-1328041" [ 984.715190] env[63293]: _type = "Task" [ 984.715190] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.722724] env[63293]: DEBUG nova.scheduler.client.report [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.735110] env[63293]: DEBUG oslo_vmware.api [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328041, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.761057] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328037, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.802134] env[63293]: DEBUG nova.objects.base [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 984.803529] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6cd440-8f6d-4b99-9c27-102bb10f12b7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.835022] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3040fe07-65ce-418d-9b8f-e9b463a93cc9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.841037] env[63293]: DEBUG oslo_vmware.api [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 984.841037] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d8ca58-9775-a81a-2ddd-fe2710440438" [ 984.841037] env[63293]: _type = "Task" [ 984.841037] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.853397] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328038, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.857741] env[63293]: DEBUG oslo_vmware.api [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d8ca58-9775-a81a-2ddd-fe2710440438, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.860362] env[63293]: INFO nova.compute.manager [-] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Took 1.38 seconds to deallocate network for instance. [ 984.898682] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.904222] env[63293]: DEBUG nova.network.neutron [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance_info_cache with network_info: [{"id": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "address": "fa:16:3e:c2:85:7d", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83a30fb0-a4", "ovs_interfaceid": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.909219] env[63293]: DEBUG oslo_vmware.api [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328039, 'name': ReconfigVM_Task, 'duration_secs': 0.457874} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.910883] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Reconfigured VM instance instance-0000003f to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 984.915048] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8ec79c7-6bdf-4ce0-a3f2-b12c0b1768d8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.932182] env[63293]: DEBUG oslo_vmware.api [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 984.932182] env[63293]: value = "task-1328042" [ 984.932182] env[63293]: _type = "Task" [ 984.932182] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.941741] env[63293]: DEBUG oslo_vmware.api [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328042, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.004280] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.004495] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.004732] env[63293]: DEBUG nova.network.neutron [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 985.004922] env[63293]: DEBUG nova.objects.instance [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'info_cache' on Instance uuid 9ddf4f23-1279-4cbd-8212-10f344060445 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.226239] env[63293]: DEBUG oslo_vmware.api [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328041, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.229111] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.787s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.252935] env[63293]: INFO nova.scheduler.client.report [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted allocations for instance 828180c0-b7f8-4666-9f5d-e6741ef6495a [ 985.259844] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328037, 'name': CreateSnapshot_Task, 'duration_secs': 0.582793} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.260361] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Created Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 985.261146] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6feabcd6-4400-4e71-adaa-4a149f8c789b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.355520] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328038, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.756455} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.358745] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2/9937a3c4-c3a6-423b-ac4f-50b7ec5879d2.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 985.358969] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.359486] env[63293]: DEBUG oslo_vmware.api [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d8ca58-9775-a81a-2ddd-fe2710440438, 'name': SearchDatastore_Task, 'duration_secs': 0.025643} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.359694] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74fcc6cd-e924-482f-8fdb-a285c1a5fae8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.361876] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.362122] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.369866] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.370248] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 985.370248] env[63293]: value = "task-1328043" [ 985.370248] env[63293]: _type = "Task" [ 985.370248] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.378827] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328043, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.395100] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.410137] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.410503] env[63293]: DEBUG nova.compute.manager [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Instance network_info: |[{"id": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "address": "fa:16:3e:c2:85:7d", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83a30fb0-a4", "ovs_interfaceid": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 985.410819] env[63293]: DEBUG oslo_concurrency.lockutils [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] Acquired lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.411036] env[63293]: DEBUG nova.network.neutron [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Refreshing network info cache for port 83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.412204] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:85:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.419693] env[63293]: DEBUG oslo.service.loopingcall [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.420556] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.420777] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-309f4f73-0df4-408a-9211-0a18a825175e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.444364] env[63293]: DEBUG oslo_vmware.api [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328042, 'name': ReconfigVM_Task, 'duration_secs': 0.234693} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.445663] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283793', 'volume_id': '658fbd49-f54e-404a-99c2-e7147e4a3700', 'name': 'volume-658fbd49-f54e-404a-99c2-e7147e4a3700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'ad585ebb-2072-45df-b645-94c9fa93576b', 'attached_at': '', 'detached_at': '', 'volume_id': '658fbd49-f54e-404a-99c2-e7147e4a3700', 'serial': '658fbd49-f54e-404a-99c2-e7147e4a3700'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 985.447739] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.447739] env[63293]: value = "task-1328044" [ 985.447739] env[63293]: _type = "Task" [ 985.447739] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.457185] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328044, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.509122] env[63293]: DEBUG nova.objects.base [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Object Instance<9ddf4f23-1279-4cbd-8212-10f344060445> lazy-loaded attributes: flavor,info_cache {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 985.727019] env[63293]: DEBUG oslo_vmware.api [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328041, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.979421} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.727385] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.727727] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.728091] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.728415] env[63293]: INFO nova.compute.manager [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Took 1.72 seconds to destroy the instance on the hypervisor. [ 985.728819] env[63293]: DEBUG oslo.service.loopingcall [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.729102] env[63293]: DEBUG nova.compute.manager [-] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 985.729231] env[63293]: DEBUG nova.network.neutron [-] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 985.765816] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d0f21ad6-b61c-4b73-aa06-fb259392584e tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "828180c0-b7f8-4666-9f5d-e6741ef6495a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.164s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.778177] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Creating linked-clone VM from snapshot {{(pid=63293) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 985.778825] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-78f85801-e1e5-47d1-aa82-4c1b7abe5cae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.788528] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 985.788528] env[63293]: value = "task-1328045" [ 985.788528] env[63293]: _type = "Task" [ 985.788528] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.798432] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328045, 'name': CloneVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.881181] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328043, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.292814} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.881421] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.882315] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8d231d-e6a3-4be9-a4c6-4a90c409bc77 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.911636] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2/9937a3c4-c3a6-423b-ac4f-50b7ec5879d2.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.919076] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0ac48d5-0d34-4bce-87fa-26b4731efd80 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.945201] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.953472] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 985.953472] env[63293]: value = "task-1328046" [ 985.953472] env[63293]: _type = "Task" [ 985.953472] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.964312] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328044, 'name': CreateVM_Task, 'duration_secs': 0.454545} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.967098] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.967565] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328046, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.970665] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.970866] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.971242] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.971725] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99ae97e4-d339-49a5-bf1c-0b7dbc004e46 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.979311] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 985.979311] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5213d652-1ae8-1e47-57c5-c95e2357743c" [ 985.979311] env[63293]: _type = "Task" [ 985.979311] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.992293] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5213d652-1ae8-1e47-57c5-c95e2357743c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.001561] env[63293]: DEBUG nova.objects.instance [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lazy-loading 'flavor' on Instance uuid ad585ebb-2072-45df-b645-94c9fa93576b {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.102724] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f67d840-80bb-405c-a926-8e1e62c5a753 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.112572] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc71d8f8-a042-40e4-827d-5a9cb1167c2e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.150668] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829a7c7a-1f04-4d1e-922a-bc7b001bf76c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.163694] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49efc78c-4079-4c34-bed8-e87f19733f05 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.181635] env[63293]: DEBUG nova.compute.provider_tree [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.301288] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328045, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.400314] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.456284] env[63293]: DEBUG nova.network.neutron [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updated VIF entry in instance network info cache for port 83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 986.456666] env[63293]: DEBUG nova.network.neutron [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance_info_cache with network_info: [{"id": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "address": "fa:16:3e:c2:85:7d", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83a30fb0-a4", "ovs_interfaceid": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.472545] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328046, 'name': ReconfigVM_Task, 'duration_secs': 0.290919} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.473678] env[63293]: DEBUG nova.network.neutron [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance_info_cache with network_info: [{"id": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "address": "fa:16:3e:93:d3:25", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dbccf6a-05", "ovs_interfaceid": "4dbccf6a-053e-4933-91bc-1b7af5a260be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.475478] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2/9937a3c4-c3a6-423b-ac4f-50b7ec5879d2.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.476348] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-21929b78-54ea-470c-943e-ee31598cec04 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.485553] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 986.485553] env[63293]: value = "task-1328047" [ 986.485553] env[63293]: _type = "Task" [ 986.485553] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.496977] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5213d652-1ae8-1e47-57c5-c95e2357743c, 'name': SearchDatastore_Task, 'duration_secs': 0.032509} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.496977] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.496977] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.497180] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.497409] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.497845] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.498402] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02d6a9d2-5729-4c3b-a3d4-90cbae44d011 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.504281] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328047, 'name': Rename_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.514648] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.514648] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 986.517935] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0aa4419-99eb-4f2d-b998-11078f1ff55f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.522678] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 986.522678] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528f97e3-080e-2c4b-68a5-6f07faaa1744" [ 986.522678] env[63293]: _type = "Task" [ 986.522678] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.533116] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528f97e3-080e-2c4b-68a5-6f07faaa1744, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.642948] env[63293]: DEBUG nova.compute.manager [req-74d9e526-bb1f-42ca-a6ea-6b5d16764f1f req-aa022337-eadd-46ea-b71b-bfba357a6fc2 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Received event network-vif-deleted-9e8240fa-3546-43eb-9cdc-4b42859690b3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.643142] env[63293]: INFO nova.compute.manager [req-74d9e526-bb1f-42ca-a6ea-6b5d16764f1f req-aa022337-eadd-46ea-b71b-bfba357a6fc2 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Neutron deleted interface 9e8240fa-3546-43eb-9cdc-4b42859690b3; detaching it from the instance and deleting it from the info cache [ 986.643303] env[63293]: DEBUG nova.network.neutron [req-74d9e526-bb1f-42ca-a6ea-6b5d16764f1f req-aa022337-eadd-46ea-b71b-bfba357a6fc2 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.685431] env[63293]: DEBUG nova.scheduler.client.report [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.798904] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328045, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.899079] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.966814] env[63293]: DEBUG oslo_concurrency.lockutils [req-f04287c5-2e3a-4fa8-a09e-3a7a03e3173f req-ea6e1152-809a-450a-be90-d7dd597e4064 service nova] Releasing lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.978337] env[63293]: DEBUG oslo_concurrency.lockutils [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-9ddf4f23-1279-4cbd-8212-10f344060445" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.998193] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328047, 'name': Rename_Task, 'duration_secs': 0.150334} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.998193] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 986.998193] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-841aab5f-4871-4786-aa54-2daed5dd08a9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.004782] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 987.004782] env[63293]: value = "task-1328048" [ 987.004782] env[63293]: _type = "Task" [ 987.004782] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.012893] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328048, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.013317] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec0fb8df-0152-46d2-b4cf-35dba110713a tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.293s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.035923] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528f97e3-080e-2c4b-68a5-6f07faaa1744, 'name': SearchDatastore_Task, 'duration_secs': 0.056862} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.036965] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9096a66e-82ca-41e5-a0cd-c7980531c7b7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.040009] env[63293]: DEBUG nova.network.neutron [-] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.045968] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 987.045968] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267baed-86c8-0574-088e-1935001033b7" [ 987.045968] env[63293]: _type = "Task" [ 987.045968] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.055701] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267baed-86c8-0574-088e-1935001033b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.147778] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a9ffb6f3-4963-4f29-be3c-2bd102c778b2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.157454] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98dc5c0b-c4f7-4c95-995c-3d87b66a0c7b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.191413] env[63293]: DEBUG nova.compute.manager [req-74d9e526-bb1f-42ca-a6ea-6b5d16764f1f req-aa022337-eadd-46ea-b71b-bfba357a6fc2 service nova] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Detach interface failed, port_id=9e8240fa-3546-43eb-9cdc-4b42859690b3, reason: Instance 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 987.299544] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328045, 'name': CloneVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.398684] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.483795] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 987.483992] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2550e56-776b-492c-aa72-21af18c7296c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.492557] env[63293]: DEBUG oslo_vmware.api [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 987.492557] env[63293]: value = "task-1328049" [ 987.492557] env[63293]: _type = "Task" [ 987.492557] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.502185] env[63293]: DEBUG oslo_vmware.api [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328049, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.514441] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328048, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.537071] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.537317] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.542554] env[63293]: INFO nova.compute.manager [-] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Took 1.81 seconds to deallocate network for instance. [ 987.560378] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267baed-86c8-0574-088e-1935001033b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.698028] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.335s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.700694] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.331s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.700894] env[63293]: DEBUG nova.objects.instance [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lazy-loading 'resources' on Instance uuid 9d3fcf27-2c2c-4975-a802-a5c0859640d0 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.803094] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328045, 'name': CloneVM_Task, 'duration_secs': 1.531825} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.803094] env[63293]: INFO nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Created linked-clone VM from snapshot [ 987.803387] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e55245c-460d-4fa1-a179-baa9b7436b93 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.811831] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Uploading image dc0bd04e-b8f4-482d-b243-fa28e2f86738 {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 987.840107] env[63293]: DEBUG oslo_vmware.rw_handles [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 987.840107] env[63293]: value = "vm-283822" [ 987.840107] env[63293]: _type = "VirtualMachine" [ 987.840107] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 987.840495] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-fa7ca646-0fa2-4a92-ad39-fc38908f913d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.850108] env[63293]: DEBUG oslo_vmware.rw_handles [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lease: (returnval){ [ 987.850108] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52695bff-e2d6-ddb9-fc3b-867b396db720" [ 987.850108] env[63293]: _type = "HttpNfcLease" [ 987.850108] env[63293]: } obtained for exporting VM: (result){ [ 987.850108] env[63293]: value = "vm-283822" [ 987.850108] env[63293]: _type = "VirtualMachine" [ 987.850108] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 987.850515] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the lease: (returnval){ [ 987.850515] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52695bff-e2d6-ddb9-fc3b-867b396db720" [ 987.850515] env[63293]: _type = "HttpNfcLease" [ 987.850515] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 987.858554] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 987.858554] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52695bff-e2d6-ddb9-fc3b-867b396db720" [ 987.858554] env[63293]: _type = "HttpNfcLease" [ 987.858554] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 987.901343] env[63293]: DEBUG oslo_vmware.api [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328030, 'name': ReconfigVM_Task, 'duration_secs': 5.754412} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.901612] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.901823] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Reconfigured VM to detach interface {{(pid=63293) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 988.004598] env[63293]: DEBUG oslo_vmware.api [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328049, 'name': PowerOnVM_Task, 'duration_secs': 0.434988} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.004897] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 988.005114] env[63293]: DEBUG nova.compute.manager [None req-1ce09f89-90db-40dc-bfeb-b56fb7d3b2f4 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.005912] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d89d59-b347-4c41-b12d-705bc587def6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.017501] env[63293]: DEBUG oslo_vmware.api [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328048, 'name': PowerOnVM_Task, 'duration_secs': 0.861012} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.019158] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 988.019380] env[63293]: INFO nova.compute.manager [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Took 8.11 seconds to spawn the instance on the hypervisor. [ 988.019565] env[63293]: DEBUG nova.compute.manager [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.023043] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963f2028-71e7-466b-bbea-dd56a56d0474 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.040065] env[63293]: DEBUG nova.compute.manager [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 988.052471] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.060019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "ad585ebb-2072-45df-b645-94c9fa93576b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.060019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.060019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.060019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.060019] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.060019] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267baed-86c8-0574-088e-1935001033b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.060019] env[63293]: INFO nova.compute.manager [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Terminating instance [ 988.061890] env[63293]: DEBUG nova.compute.manager [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.062217] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 988.063079] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f5c821-e93b-4808-82a8-e42ea27d18d1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.071424] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.071808] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbee2656-f57a-48e6-bd8f-065769808bc7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.078936] env[63293]: DEBUG oslo_vmware.api [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 988.078936] env[63293]: value = "task-1328051" [ 988.078936] env[63293]: _type = "Task" [ 988.078936] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.091024] env[63293]: DEBUG oslo_vmware.api [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328051, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.270031] env[63293]: INFO nova.scheduler.client.report [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted allocation for migration a719d2d5-2903-4e76-9d8c-cc2588bcda26 [ 988.363017] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 988.363017] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52695bff-e2d6-ddb9-fc3b-867b396db720" [ 988.363017] env[63293]: _type = "HttpNfcLease" [ 988.363017] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 988.363343] env[63293]: DEBUG oslo_vmware.rw_handles [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 988.363343] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52695bff-e2d6-ddb9-fc3b-867b396db720" [ 988.363343] env[63293]: _type = "HttpNfcLease" [ 988.363343] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 988.364103] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27134d61-f4c0-4ebe-99f4-ee7120740bdf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.377898] env[63293]: DEBUG oslo_vmware.rw_handles [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525ffb92-802d-af5a-f761-9012be672d2b/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 988.378134] env[63293]: DEBUG oslo_vmware.rw_handles [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525ffb92-802d-af5a-f761-9012be672d2b/disk-0.vmdk for reading. {{(pid=63293) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 988.472548] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-361b7b96-479d-482e-842c-65e71ff44087 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.478730] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9605d547-0e69-468b-b152-be2754a7dd6a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.488138] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce3d51b3-2e46-41fe-bfac-4c047cc79fa0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.523448] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8059f51-82ab-4bb0-bd8d-c694050d8db9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.547627] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6ba026-827f-4324-95a9-c1aba5461f37 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.563062] env[63293]: INFO nova.compute.manager [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Took 15.71 seconds to build instance. [ 988.574039] env[63293]: DEBUG nova.compute.provider_tree [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.578608] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267baed-86c8-0574-088e-1935001033b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.591454] env[63293]: DEBUG oslo_vmware.api [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328051, 'name': PowerOffVM_Task, 'duration_secs': 0.421077} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.592574] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.596017] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 988.596017] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 988.596017] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d558716-8efb-4e68-a095-c3d2955f9827 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.777792] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2578f2ca-d6b1-45df-b0d9-fe4c869a6252 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.930s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.067633] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ada4a51f-42f7-46cc-9312-8122ba1ba3bb tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.226s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.075026] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267baed-86c8-0574-088e-1935001033b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.081671] env[63293]: DEBUG nova.scheduler.client.report [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 989.258798] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.258798] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.258798] env[63293]: DEBUG nova.network.neutron [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.273387] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.275976] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.275976] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.275976] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.275976] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.279812] env[63293]: INFO nova.compute.manager [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Terminating instance [ 989.283032] env[63293]: DEBUG nova.compute.manager [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 989.283581] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 989.285366] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac91ad6-c11b-4068-b644-a3160f722b20 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.295025] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 989.295887] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c39cb5d8-4cd9-4272-907f-1c571d817f84 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.305670] env[63293]: DEBUG oslo_vmware.api [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 989.305670] env[63293]: value = "task-1328053" [ 989.305670] env[63293]: _type = "Task" [ 989.305670] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.317072] env[63293]: DEBUG oslo_vmware.api [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328053, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.373283] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "9ddf4f23-1279-4cbd-8212-10f344060445" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.373655] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.376688] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "9ddf4f23-1279-4cbd-8212-10f344060445-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.376688] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.376688] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.380026] env[63293]: INFO nova.compute.manager [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Terminating instance [ 989.384551] env[63293]: DEBUG nova.compute.manager [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 989.386304] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 989.387735] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1164511-dbb5-46d5-8535-35769bf436d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.400134] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 989.400775] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a11caaac-0472-4881-9731-e79a1f67dc51 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.411691] env[63293]: DEBUG oslo_vmware.api [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 989.411691] env[63293]: value = "task-1328054" [ 989.411691] env[63293]: _type = "Task" [ 989.411691] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.421451] env[63293]: DEBUG oslo_vmware.api [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328054, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.507857] env[63293]: DEBUG nova.compute.manager [req-0f68af42-8aaf-4d4e-93dd-83fc69c8fcae req-9ec7b45f-0771-49ba-b602-7a623a0f0220 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Received event network-changed-df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.510657] env[63293]: DEBUG nova.compute.manager [req-0f68af42-8aaf-4d4e-93dd-83fc69c8fcae req-9ec7b45f-0771-49ba-b602-7a623a0f0220 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing instance network info cache due to event network-changed-df1c83c6-23d7-439b-a496-234c48738d81. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.510657] env[63293]: DEBUG oslo_concurrency.lockutils [req-0f68af42-8aaf-4d4e-93dd-83fc69c8fcae req-9ec7b45f-0771-49ba-b602-7a623a0f0220 service nova] Acquiring lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.570855] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267baed-86c8-0574-088e-1935001033b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.585737] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.885s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.588791] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.536s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.589785] env[63293]: DEBUG nova.objects.instance [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lazy-loading 'resources' on Instance uuid 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.620914] env[63293]: INFO nova.scheduler.client.report [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Deleted allocations for instance 9d3fcf27-2c2c-4975-a802-a5c0859640d0 [ 989.819206] env[63293]: DEBUG oslo_vmware.api [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328053, 'name': PowerOffVM_Task, 'duration_secs': 0.243712} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.819594] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 989.819706] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 989.819963] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b418e729-8a32-4327-9ef7-641734845e6c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.888315] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 989.888602] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 989.888841] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleting the datastore file [datastore1] ad585ebb-2072-45df-b645-94c9fa93576b {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.889331] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62c091c9-d736-4e8e-81fe-a18d9490ea66 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.898540] env[63293]: DEBUG oslo_vmware.api [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 989.898540] env[63293]: value = "task-1328056" [ 989.898540] env[63293]: _type = "Task" [ 989.898540] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.905701] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 989.906142] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 989.906343] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleting the datastore file [datastore2] f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.906999] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-69f70ddc-e3e1-4dee-babd-d046c146d4e9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.912805] env[63293]: DEBUG oslo_vmware.api [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.917366] env[63293]: DEBUG oslo_vmware.api [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 989.917366] env[63293]: value = "task-1328057" [ 989.917366] env[63293]: _type = "Task" [ 989.917366] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.924544] env[63293]: DEBUG oslo_vmware.api [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328054, 'name': PowerOffVM_Task, 'duration_secs': 0.215123} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.925189] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 989.925389] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 989.925766] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-809ee4d0-f019-497f-b04b-040d564015f5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.930488] env[63293]: DEBUG oslo_vmware.api [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328057, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.975678] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "71154ca4-ea2c-4187-8319-0bfb9b289eab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.975995] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "71154ca4-ea2c-4187-8319-0bfb9b289eab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.017250] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 990.017250] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 990.017250] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleting the datastore file [datastore2] 9ddf4f23-1279-4cbd-8212-10f344060445 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 990.017250] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c95c26d8-6432-4f7b-8e5e-d151ae309cf7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.024798] env[63293]: DEBUG oslo_vmware.api [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 990.024798] env[63293]: value = "task-1328059" [ 990.024798] env[63293]: _type = "Task" [ 990.024798] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.034276] env[63293]: DEBUG oslo_vmware.api [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328059, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.068598] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267baed-86c8-0574-088e-1935001033b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.130998] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33f96cb5-0ae2-43ef-99b2-e5ab1337ce4e tempest-ImagesTestJSON-1563158367 tempest-ImagesTestJSON-1563158367-project-member] Lock "9d3fcf27-2c2c-4975-a802-a5c0859640d0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.786s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.187451] env[63293]: DEBUG nova.compute.manager [req-000d042a-f5ba-4fb9-a480-7e7e5fcb6202 req-aa91406d-3e5b-4875-9174-d7859d252d8c service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Received event network-changed-cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.187892] env[63293]: DEBUG nova.compute.manager [req-000d042a-f5ba-4fb9-a480-7e7e5fcb6202 req-aa91406d-3e5b-4875-9174-d7859d252d8c service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Refreshing instance network info cache due to event network-changed-cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 990.188062] env[63293]: DEBUG oslo_concurrency.lockutils [req-000d042a-f5ba-4fb9-a480-7e7e5fcb6202 req-aa91406d-3e5b-4875-9174-d7859d252d8c service nova] Acquiring lock "refresh_cache-9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.188315] env[63293]: DEBUG oslo_concurrency.lockutils [req-000d042a-f5ba-4fb9-a480-7e7e5fcb6202 req-aa91406d-3e5b-4875-9174-d7859d252d8c service nova] Acquired lock "refresh_cache-9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.188440] env[63293]: DEBUG nova.network.neutron [req-000d042a-f5ba-4fb9-a480-7e7e5fcb6202 req-aa91406d-3e5b-4875-9174-d7859d252d8c service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Refreshing network info cache for port cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 990.190894] env[63293]: INFO nova.network.neutron [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Port c2fa0388-8ddf-497f-8200-903c0fde5a7b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 990.193879] env[63293]: DEBUG nova.network.neutron [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updating instance_info_cache with network_info: [{"id": "df1c83c6-23d7-439b-a496-234c48738d81", "address": "fa:16:3e:0f:9f:14", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1c83c6-23", "ovs_interfaceid": "df1c83c6-23d7-439b-a496-234c48738d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.244430] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "interface-f9ccb216-1435-44c5-ab34-a6388d794551-c2fa0388-8ddf-497f-8200-903c0fde5a7b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.244961] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-f9ccb216-1435-44c5-ab34-a6388d794551-c2fa0388-8ddf-497f-8200-903c0fde5a7b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.246139] env[63293]: DEBUG nova.objects.instance [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'flavor' on Instance uuid f9ccb216-1435-44c5-ab34-a6388d794551 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.315969] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6252328d-4b85-4f35-b26f-060f30e974d4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.325253] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11996af1-19c1-4dca-9044-9c6011ee5e6e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.366619] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d11c77-54fe-484a-89a1-e4a0450b47d4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.379174] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11aaab6d-8a08-4961-a27a-50c8aa2b648d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.398115] env[63293]: DEBUG nova.compute.provider_tree [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.414313] env[63293]: DEBUG oslo_vmware.api [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.428514] env[63293]: DEBUG oslo_vmware.api [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328057, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.340828} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.429059] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.429398] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.429696] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.429994] env[63293]: INFO nova.compute.manager [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Took 1.15 seconds to destroy the instance on the hypervisor. [ 990.430392] env[63293]: DEBUG oslo.service.loopingcall [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.430989] env[63293]: DEBUG nova.compute.manager [-] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 990.431222] env[63293]: DEBUG nova.network.neutron [-] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 990.482569] env[63293]: DEBUG nova.compute.manager [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 990.538612] env[63293]: DEBUG oslo_vmware.api [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328059, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.330215} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.539644] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.540152] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.540935] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.543015] env[63293]: INFO nova.compute.manager [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Took 1.16 seconds to destroy the instance on the hypervisor. [ 990.543015] env[63293]: DEBUG oslo.service.loopingcall [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.543015] env[63293]: DEBUG nova.compute.manager [-] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 990.543015] env[63293]: DEBUG nova.network.neutron [-] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 990.571241] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267baed-86c8-0574-088e-1935001033b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.701689] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.705715] env[63293]: DEBUG oslo_concurrency.lockutils [req-0f68af42-8aaf-4d4e-93dd-83fc69c8fcae req-9ec7b45f-0771-49ba-b602-7a623a0f0220 service nova] Acquired lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.706025] env[63293]: DEBUG nova.network.neutron [req-0f68af42-8aaf-4d4e-93dd-83fc69c8fcae req-9ec7b45f-0771-49ba-b602-7a623a0f0220 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Refreshing network info cache for port df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 990.911332] env[63293]: DEBUG nova.scheduler.client.report [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.916094] env[63293]: DEBUG oslo_vmware.api [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.010796] env[63293]: DEBUG nova.objects.instance [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'pci_requests' on Instance uuid f9ccb216-1435-44c5-ab34-a6388d794551 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.011901] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.070530] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267baed-86c8-0574-088e-1935001033b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.078994] env[63293]: DEBUG nova.network.neutron [req-000d042a-f5ba-4fb9-a480-7e7e5fcb6202 req-aa91406d-3e5b-4875-9174-d7859d252d8c service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Updated VIF entry in instance network info cache for port cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.079129] env[63293]: DEBUG nova.network.neutron [req-000d042a-f5ba-4fb9-a480-7e7e5fcb6202 req-aa91406d-3e5b-4875-9174-d7859d252d8c service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Updating instance_info_cache with network_info: [{"id": "cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3", "address": "fa:16:3e:24:8c:84", "network": {"id": "9d5a3e58-567a-4c3f-9ad9-9fcbf81fced9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-949261114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.235", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1a30d37c1b47fd8670db0a9ef7020d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc516442-7d", "ovs_interfaceid": "cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.207359] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7dbf3d21-5fdb-4a29-80ed-17bb8da12961 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-35cbf277-b85a-4101-beae-0125fb5a61f5-c2fa0388-8ddf-497f-8200-903c0fde5a7b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.935s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.208050] env[63293]: DEBUG nova.network.neutron [-] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.411170] env[63293]: DEBUG oslo_vmware.api [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.418144] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.829s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.421392] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.828s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.422337] env[63293]: INFO nova.compute.claims [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 991.460900] env[63293]: INFO nova.scheduler.client.report [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleted allocations for instance 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535 [ 991.513174] env[63293]: DEBUG nova.objects.base [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 991.513463] env[63293]: DEBUG nova.network.neutron [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 991.539799] env[63293]: DEBUG nova.compute.manager [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Received event network-changed-5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.539799] env[63293]: DEBUG nova.compute.manager [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Refreshing instance network info cache due to event network-changed-5d29fa7c-6329-4700-b988-ac08592d4df7. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.539799] env[63293]: DEBUG oslo_concurrency.lockutils [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] Acquiring lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.541049] env[63293]: DEBUG oslo_concurrency.lockutils [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] Acquired lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.541049] env[63293]: DEBUG nova.network.neutron [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Refreshing network info cache for port 5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.559653] env[63293]: DEBUG nova.network.neutron [req-0f68af42-8aaf-4d4e-93dd-83fc69c8fcae req-9ec7b45f-0771-49ba-b602-7a623a0f0220 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updated VIF entry in instance network info cache for port df1c83c6-23d7-439b-a496-234c48738d81. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.560178] env[63293]: DEBUG nova.network.neutron [req-0f68af42-8aaf-4d4e-93dd-83fc69c8fcae req-9ec7b45f-0771-49ba-b602-7a623a0f0220 service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updating instance_info_cache with network_info: [{"id": "df1c83c6-23d7-439b-a496-234c48738d81", "address": "fa:16:3e:0f:9f:14", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1c83c6-23", "ovs_interfaceid": "df1c83c6-23d7-439b-a496-234c48738d81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.573972] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267baed-86c8-0574-088e-1935001033b7, 'name': SearchDatastore_Task, 'duration_secs': 4.514203} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.573972] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.574061] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] d605e066-806a-4aac-9af6-f79d542ab2e3/d605e066-806a-4aac-9af6-f79d542ab2e3.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.574572] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2dfdade-ba73-4564-b214-2c55ad6bb009 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.582247] env[63293]: DEBUG oslo_concurrency.lockutils [req-000d042a-f5ba-4fb9-a480-7e7e5fcb6202 req-aa91406d-3e5b-4875-9174-d7859d252d8c service nova] Releasing lock "refresh_cache-9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.584462] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 991.584462] env[63293]: value = "task-1328060" [ 991.584462] env[63293]: _type = "Task" [ 991.584462] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.593729] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328060, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.644179] env[63293]: DEBUG nova.policy [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2312d48038d7474481a0709267645780', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '866b347100454019b07f63922b995bb6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.713614] env[63293]: INFO nova.compute.manager [-] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Took 1.28 seconds to deallocate network for instance. [ 991.742486] env[63293]: DEBUG nova.network.neutron [-] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.913190] env[63293]: DEBUG oslo_vmware.api [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.970863] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f7175b6d-4937-416d-a553-7a2b7ffdddfe tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "2f55ac2a-63d1-4713-ab0c-47fb6cf7d535" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.966s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.063412] env[63293]: DEBUG oslo_concurrency.lockutils [req-0f68af42-8aaf-4d4e-93dd-83fc69c8fcae req-9ec7b45f-0771-49ba-b602-7a623a0f0220 service nova] Releasing lock "refresh_cache-35cbf277-b85a-4101-beae-0125fb5a61f5" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.098292] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328060, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.216903] env[63293]: DEBUG nova.compute.manager [req-59ed1e0c-f7d0-4440-8429-f825ee5a66a4 req-b11f7057-8756-4b8f-a08a-6617afdaef94 service nova] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Received event network-vif-deleted-a20688c6-3fbe-4355-85f9-334019e37f43 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.223949] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.248760] env[63293]: INFO nova.compute.manager [-] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Took 1.71 seconds to deallocate network for instance. [ 992.416683] env[63293]: DEBUG oslo_vmware.api [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328056, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.097031} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.416990] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.417231] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 992.417444] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 992.417648] env[63293]: INFO nova.compute.manager [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Took 4.36 seconds to destroy the instance on the hypervisor. [ 992.417905] env[63293]: DEBUG oslo.service.loopingcall [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.418143] env[63293]: DEBUG nova.compute.manager [-] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.418257] env[63293]: DEBUG nova.network.neutron [-] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 992.600821] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328060, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.9803} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.600821] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] d605e066-806a-4aac-9af6-f79d542ab2e3/d605e066-806a-4aac-9af6-f79d542ab2e3.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 992.600821] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.600821] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63705634-8641-49eb-b129-16bcae5f8601 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.609325] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 992.609325] env[63293]: value = "task-1328061" [ 992.609325] env[63293]: _type = "Task" [ 992.609325] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.626371] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328061, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.636672] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c2ec26-528a-4edd-aa85-a3edbb2cad64 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.647773] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d184714a-1fe6-4d1a-a507-fb50b41805b7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.687134] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0e7c59-a99f-4efc-a797-fa845de90c08 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.699025] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4afa33-1b2a-4ac3-a5bb-4c15d1e88375 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.716359] env[63293]: DEBUG nova.compute.provider_tree [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.759369] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.767830] env[63293]: DEBUG nova.network.neutron [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updated VIF entry in instance network info cache for port 5d29fa7c-6329-4700-b988-ac08592d4df7. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.768260] env[63293]: DEBUG nova.network.neutron [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updating instance_info_cache with network_info: [{"id": "5d29fa7c-6329-4700-b988-ac08592d4df7", "address": "fa:16:3e:56:b9:c4", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d29fa7c-63", "ovs_interfaceid": "5d29fa7c-6329-4700-b988-ac08592d4df7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.120299] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328061, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120145} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.120597] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 993.121497] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb0d1ffb-f757-434f-8218-03089deacce9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.151049] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] d605e066-806a-4aac-9af6-f79d542ab2e3/d605e066-806a-4aac-9af6-f79d542ab2e3.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 993.151049] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3342d62a-f4d3-48a9-978f-940bebc6e088 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.179738] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 993.179738] env[63293]: value = "task-1328062" [ 993.179738] env[63293]: _type = "Task" [ 993.179738] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.189510] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328062, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.222545] env[63293]: DEBUG nova.scheduler.client.report [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.271439] env[63293]: DEBUG oslo_concurrency.lockutils [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] Releasing lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.271749] env[63293]: DEBUG nova.compute.manager [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Received event network-vif-deleted-4dbccf6a-053e-4933-91bc-1b7af5a260be {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.271934] env[63293]: INFO nova.compute.manager [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Neutron deleted interface 4dbccf6a-053e-4933-91bc-1b7af5a260be; detaching it from the instance and deleting it from the info cache [ 993.272152] env[63293]: DEBUG nova.network.neutron [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.440565] env[63293]: DEBUG nova.network.neutron [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Successfully updated port: c2fa0388-8ddf-497f-8200-903c0fde5a7b {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.547569] env[63293]: DEBUG nova.network.neutron [-] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.562395] env[63293]: DEBUG oslo_concurrency.lockutils [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "d753598d-a92a-4515-9ad1-d386294f7a99" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.563183] env[63293]: DEBUG oslo_concurrency.lockutils [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.587770] env[63293]: DEBUG nova.compute.manager [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Received event network-vif-deleted-0576b708-1b9b-4159-8079-9b97d8b4e70d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.588082] env[63293]: DEBUG nova.compute.manager [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Received event network-vif-plugged-c2fa0388-8ddf-497f-8200-903c0fde5a7b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.588334] env[63293]: DEBUG oslo_concurrency.lockutils [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] Acquiring lock "f9ccb216-1435-44c5-ab34-a6388d794551-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.588581] env[63293]: DEBUG oslo_concurrency.lockutils [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] Lock "f9ccb216-1435-44c5-ab34-a6388d794551-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.588773] env[63293]: DEBUG oslo_concurrency.lockutils [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] Lock "f9ccb216-1435-44c5-ab34-a6388d794551-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.588973] env[63293]: DEBUG nova.compute.manager [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] No waiting events found dispatching network-vif-plugged-c2fa0388-8ddf-497f-8200-903c0fde5a7b {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.589208] env[63293]: WARNING nova.compute.manager [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Received unexpected event network-vif-plugged-c2fa0388-8ddf-497f-8200-903c0fde5a7b for instance with vm_state active and task_state None. [ 993.589429] env[63293]: DEBUG nova.compute.manager [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Received event network-changed-c2fa0388-8ddf-497f-8200-903c0fde5a7b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.589600] env[63293]: DEBUG nova.compute.manager [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Refreshing instance network info cache due to event network-changed-c2fa0388-8ddf-497f-8200-903c0fde5a7b. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 993.589837] env[63293]: DEBUG oslo_concurrency.lockutils [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] Acquiring lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.590016] env[63293]: DEBUG oslo_concurrency.lockutils [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] Acquired lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.590246] env[63293]: DEBUG nova.network.neutron [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Refreshing network info cache for port c2fa0388-8ddf-497f-8200-903c0fde5a7b {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.693242] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328062, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.729130] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.309s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.729710] env[63293]: DEBUG nova.compute.manager [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 993.732500] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.721s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.733968] env[63293]: INFO nova.compute.claims [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.774954] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f5b6fd0-cfe8-4e53-a824-4a0e9bf5e1b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.786776] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e441f5ec-417c-4b18-9e2e-3b4408fb3d58 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.822599] env[63293]: DEBUG nova.compute.manager [req-06ca78de-25dc-412b-b978-3a4192614042 req-0603c0e2-c42d-451e-b4c6-c2707fd1eda7 service nova] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Detach interface failed, port_id=4dbccf6a-053e-4933-91bc-1b7af5a260be, reason: Instance 9ddf4f23-1279-4cbd-8212-10f344060445 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 993.943933] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.050815] env[63293]: INFO nova.compute.manager [-] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Took 1.63 seconds to deallocate network for instance. [ 994.066219] env[63293]: INFO nova.compute.manager [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Detaching volume 6c9dda44-62d5-4ee9-8961-4b9fec04802a [ 994.104025] env[63293]: INFO nova.virt.block_device [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Attempting to driver detach volume 6c9dda44-62d5-4ee9-8961-4b9fec04802a from mountpoint /dev/sdb [ 994.104282] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Volume detach. Driver type: vmdk {{(pid=63293) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 994.104528] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283791', 'volume_id': '6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'name': 'volume-6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd753598d-a92a-4515-9ad1-d386294f7a99', 'attached_at': '', 'detached_at': '', 'volume_id': '6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'serial': '6c9dda44-62d5-4ee9-8961-4b9fec04802a'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 994.105388] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc97970-2803-4aed-abed-f3d3ca619c3e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.131531] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c0d9d6-54cf-49e1-877d-3966114db885 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.141356] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad65312c-b1b0-4137-ba29-b9690df99877 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.163430] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a498b21e-ec97-495a-99e1-426e683ed4a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.180015] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] The volume has not been displaced from its original location: [datastore1] volume-6c9dda44-62d5-4ee9-8961-4b9fec04802a/volume-6c9dda44-62d5-4ee9-8961-4b9fec04802a.vmdk. No consolidation needed. {{(pid=63293) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 994.185097] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Reconfiguring VM instance instance-0000004a to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 994.187873] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-009056e0-bc97-4262-b666-a3439b056475 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.210383] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328062, 'name': ReconfigVM_Task, 'duration_secs': 0.570083} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.211822] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Reconfigured VM instance instance-00000060 to attach disk [datastore1] d605e066-806a-4aac-9af6-f79d542ab2e3/d605e066-806a-4aac-9af6-f79d542ab2e3.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.212620] env[63293]: DEBUG oslo_vmware.api [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 994.212620] env[63293]: value = "task-1328063" [ 994.212620] env[63293]: _type = "Task" [ 994.212620] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.212846] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e74a26ba-f560-455a-a681-aa920b5a20b6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.226723] env[63293]: DEBUG oslo_vmware.api [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328063, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.228559] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 994.228559] env[63293]: value = "task-1328064" [ 994.228559] env[63293]: _type = "Task" [ 994.228559] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.240627] env[63293]: DEBUG nova.compute.utils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 994.244029] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328064, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.244398] env[63293]: DEBUG nova.compute.manager [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 994.244978] env[63293]: DEBUG nova.network.neutron [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 994.334020] env[63293]: DEBUG nova.policy [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b94ec4d2961a43ca9a331a6051fa774d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7744835de0bc4e30b1b29ee5a439cd0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 994.557797] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.716889] env[63293]: DEBUG nova.network.neutron [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Added VIF to instance network info cache for port c2fa0388-8ddf-497f-8200-903c0fde5a7b. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 994.717461] env[63293]: DEBUG nova.network.neutron [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updating instance_info_cache with network_info: [{"id": "5d29fa7c-6329-4700-b988-ac08592d4df7", "address": "fa:16:3e:56:b9:c4", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d29fa7c-63", "ovs_interfaceid": "5d29fa7c-6329-4700-b988-ac08592d4df7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2fa0388-8ddf-497f-8200-903c0fde5a7b", "address": "fa:16:3e:ae:52:3b", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2fa0388-8d", "ovs_interfaceid": "c2fa0388-8ddf-497f-8200-903c0fde5a7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.728928] env[63293]: DEBUG oslo_vmware.api [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328063, 'name': ReconfigVM_Task, 'duration_secs': 0.297551} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.729556] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Reconfigured VM instance instance-0000004a to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 994.737708] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-522c556b-59bd-4e28-a6f5-9d2532c8133a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.751403] env[63293]: DEBUG nova.compute.manager [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 994.766249] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328064, 'name': Rename_Task, 'duration_secs': 0.204712} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.767871] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.768538] env[63293]: DEBUG oslo_vmware.api [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 994.768538] env[63293]: value = "task-1328065" [ 994.768538] env[63293]: _type = "Task" [ 994.768538] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.768735] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b9b317ef-339b-46d8-aab5-6797b07518f7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.781991] env[63293]: DEBUG oslo_vmware.api [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328065, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.787711] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 994.787711] env[63293]: value = "task-1328066" [ 994.787711] env[63293]: _type = "Task" [ 994.787711] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.798027] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328066, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.954232] env[63293]: DEBUG nova.network.neutron [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Successfully created port: 12ae71a2-a302-4223-8d65-630cde456181 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 995.009918] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c318ea0-4c40-4fed-92d9-ad37658ef54e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.018739] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4e7df7-700d-4066-8b8f-6420e4f77271 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.054423] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e567b69e-ecfe-4ff3-83e8-0814cec9f66f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.063679] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01f1ce6-d88f-4638-b663-5a84326a8cbe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.081734] env[63293]: DEBUG nova.compute.provider_tree [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.223731] env[63293]: DEBUG oslo_concurrency.lockutils [req-a61c5441-5dc2-4a1a-83ab-903abdc6aca3 req-0c6253e6-c64c-420d-9f3a-9ada26214fb6 service nova] Releasing lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.224445] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.224754] env[63293]: DEBUG nova.network.neutron [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 995.281658] env[63293]: DEBUG oslo_vmware.api [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328065, 'name': ReconfigVM_Task, 'duration_secs': 0.19358} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.282164] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283791', 'volume_id': '6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'name': 'volume-6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd753598d-a92a-4515-9ad1-d386294f7a99', 'attached_at': '', 'detached_at': '', 'volume_id': '6c9dda44-62d5-4ee9-8961-4b9fec04802a', 'serial': '6c9dda44-62d5-4ee9-8961-4b9fec04802a'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 995.297744] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328066, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.588187] env[63293]: DEBUG nova.scheduler.client.report [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.763191] env[63293]: DEBUG nova.compute.manager [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 995.770652] env[63293]: WARNING nova.network.neutron [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] 908b99b3-6031-4de0-86a4-09a9cbe355a5 already exists in list: networks containing: ['908b99b3-6031-4de0-86a4-09a9cbe355a5']. ignoring it [ 995.770652] env[63293]: WARNING nova.network.neutron [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] 908b99b3-6031-4de0-86a4-09a9cbe355a5 already exists in list: networks containing: ['908b99b3-6031-4de0-86a4-09a9cbe355a5']. ignoring it [ 995.770786] env[63293]: WARNING nova.network.neutron [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] c2fa0388-8ddf-497f-8200-903c0fde5a7b already exists in list: port_ids containing: ['c2fa0388-8ddf-497f-8200-903c0fde5a7b']. ignoring it [ 995.794042] env[63293]: DEBUG nova.virt.hardware [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.794339] env[63293]: DEBUG nova.virt.hardware [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.794539] env[63293]: DEBUG nova.virt.hardware [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.794776] env[63293]: DEBUG nova.virt.hardware [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.794966] env[63293]: DEBUG nova.virt.hardware [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.795156] env[63293]: DEBUG nova.virt.hardware [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.795410] env[63293]: DEBUG nova.virt.hardware [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.796019] env[63293]: DEBUG nova.virt.hardware [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.796019] env[63293]: DEBUG nova.virt.hardware [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.796019] env[63293]: DEBUG nova.virt.hardware [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.797328] env[63293]: DEBUG nova.virt.hardware [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.797328] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b675d29-360e-4350-9646-332c223f6f29 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.809970] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3fbd6bc-3972-4719-a87d-b507de5ab385 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.816371] env[63293]: DEBUG oslo_vmware.api [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328066, 'name': PowerOnVM_Task, 'duration_secs': 0.721303} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.816882] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.817125] env[63293]: INFO nova.compute.manager [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Took 12.71 seconds to spawn the instance on the hypervisor. [ 995.818700] env[63293]: DEBUG nova.compute.manager [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.818700] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a1c128-4b22-4a64-bce2-77e90ace2ab4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.856857] env[63293]: DEBUG nova.objects.instance [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lazy-loading 'flavor' on Instance uuid d753598d-a92a-4515-9ad1-d386294f7a99 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.093836] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.094271] env[63293]: DEBUG nova.compute.manager [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 996.103725] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.877s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.103725] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.103725] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.345s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.103725] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.105770] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.548s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.105770] env[63293]: DEBUG nova.objects.instance [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lazy-loading 'resources' on Instance uuid ad585ebb-2072-45df-b645-94c9fa93576b {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.143879] env[63293]: INFO nova.scheduler.client.report [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted allocations for instance f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a [ 996.149121] env[63293]: INFO nova.scheduler.client.report [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleted allocations for instance 9ddf4f23-1279-4cbd-8212-10f344060445 [ 996.285858] env[63293]: DEBUG nova.network.neutron [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updating instance_info_cache with network_info: [{"id": "5d29fa7c-6329-4700-b988-ac08592d4df7", "address": "fa:16:3e:56:b9:c4", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d29fa7c-63", "ovs_interfaceid": "5d29fa7c-6329-4700-b988-ac08592d4df7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2fa0388-8ddf-497f-8200-903c0fde5a7b", "address": "fa:16:3e:ae:52:3b", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2fa0388-8d", "ovs_interfaceid": "c2fa0388-8ddf-497f-8200-903c0fde5a7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.350301] env[63293]: INFO nova.compute.manager [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Took 20.53 seconds to build instance. [ 996.462949] env[63293]: DEBUG oslo_vmware.rw_handles [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525ffb92-802d-af5a-f761-9012be672d2b/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 996.465065] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0695d73c-ea6f-4106-9b34-9a963204ff01 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.474120] env[63293]: DEBUG oslo_vmware.rw_handles [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525ffb92-802d-af5a-f761-9012be672d2b/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 996.474298] env[63293]: ERROR oslo_vmware.rw_handles [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525ffb92-802d-af5a-f761-9012be672d2b/disk-0.vmdk due to incomplete transfer. [ 996.474540] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6e3844cb-3772-4eb3-ac85-35ffb252f4d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.483858] env[63293]: DEBUG oslo_vmware.rw_handles [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525ffb92-802d-af5a-f761-9012be672d2b/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 996.484066] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Uploaded image dc0bd04e-b8f4-482d-b243-fa28e2f86738 to the Glance image server {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 996.486518] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Destroying the VM {{(pid=63293) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 996.488128] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3590a2e8-0f92-402f-a163-d44ab5a1a57e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.496019] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 996.496019] env[63293]: value = "task-1328067" [ 996.496019] env[63293]: _type = "Task" [ 996.496019] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.504462] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328067, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.604060] env[63293]: DEBUG nova.compute.utils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.605025] env[63293]: DEBUG nova.compute.manager [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 996.605025] env[63293]: DEBUG nova.network.neutron [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 996.643077] env[63293]: DEBUG nova.compute.manager [req-fe478608-c560-49ac-82aa-5d19ef0d79de req-7b772756-9e84-4011-bd32-3b6309908a06 service nova] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Received event network-vif-plugged-12ae71a2-a302-4223-8d65-630cde456181 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.643077] env[63293]: DEBUG oslo_concurrency.lockutils [req-fe478608-c560-49ac-82aa-5d19ef0d79de req-7b772756-9e84-4011-bd32-3b6309908a06 service nova] Acquiring lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.643077] env[63293]: DEBUG oslo_concurrency.lockutils [req-fe478608-c560-49ac-82aa-5d19ef0d79de req-7b772756-9e84-4011-bd32-3b6309908a06 service nova] Lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.643077] env[63293]: DEBUG oslo_concurrency.lockutils [req-fe478608-c560-49ac-82aa-5d19ef0d79de req-7b772756-9e84-4011-bd32-3b6309908a06 service nova] Lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.643077] env[63293]: DEBUG nova.compute.manager [req-fe478608-c560-49ac-82aa-5d19ef0d79de req-7b772756-9e84-4011-bd32-3b6309908a06 service nova] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] No waiting events found dispatching network-vif-plugged-12ae71a2-a302-4223-8d65-630cde456181 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 996.643487] env[63293]: WARNING nova.compute.manager [req-fe478608-c560-49ac-82aa-5d19ef0d79de req-7b772756-9e84-4011-bd32-3b6309908a06 service nova] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Received unexpected event network-vif-plugged-12ae71a2-a302-4223-8d65-630cde456181 for instance with vm_state building and task_state spawning. [ 996.651345] env[63293]: DEBUG oslo_concurrency.lockutils [None req-9a06286a-c589-4032-b68b-d0675c7918fe tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.377s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.661461] env[63293]: DEBUG oslo_concurrency.lockutils [None req-2a0e9a8b-0a97-4fa4-8229-8aee8b584b0a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "9ddf4f23-1279-4cbd-8212-10f344060445" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.288s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.700222] env[63293]: DEBUG nova.policy [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6860a9e34a1b44029cf57d9f04ae87c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fac34e49d5dc49e7a7055c998d5b6766', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 996.780149] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfabbb8b-3918-45cd-9545-84e50debf1d8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.788607] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.790091] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.791266] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.792147] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b649147-d177-46ef-aafc-b3f4d16e0d69 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.797045] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b32828-d103-4e28-87f6-acc2ba175b37 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.802868] env[63293]: DEBUG nova.network.neutron [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Successfully updated port: 12ae71a2-a302-4223-8d65-630cde456181 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 996.848614] env[63293]: DEBUG nova.virt.hardware [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 996.848932] env[63293]: DEBUG nova.virt.hardware [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 996.849143] env[63293]: DEBUG nova.virt.hardware [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 996.849344] env[63293]: DEBUG nova.virt.hardware [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 996.849500] env[63293]: DEBUG nova.virt.hardware [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 996.849654] env[63293]: DEBUG nova.virt.hardware [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 996.849866] env[63293]: DEBUG nova.virt.hardware [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 996.850050] env[63293]: DEBUG nova.virt.hardware [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 996.851379] env[63293]: DEBUG nova.virt.hardware [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 996.851379] env[63293]: DEBUG nova.virt.hardware [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 996.851379] env[63293]: DEBUG nova.virt.hardware [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 996.857225] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Reconfiguring VM to attach interface {{(pid=63293) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 996.858365] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d6f7c3f1-8314-491f-b37b-6cf724fcd196 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.048s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.858570] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e116660-a7e0-4b76-bdb4-ddc4ea635294 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.871981] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74fe44f-ecc0-439d-8e0a-9c822e8a6cb5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.875163] env[63293]: DEBUG oslo_concurrency.lockutils [None req-47bf4b10-2f6d-4ac6-82a4-f2dbb1954bee tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.312s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.883617] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1b3fb4-c1fd-4971-a477-0111f157efa1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.888140] env[63293]: DEBUG oslo_vmware.api [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 996.888140] env[63293]: value = "task-1328068" [ 996.888140] env[63293]: _type = "Task" [ 996.888140] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.904849] env[63293]: DEBUG nova.compute.provider_tree [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.913276] env[63293]: DEBUG oslo_vmware.api [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328068, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.005473] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328067, 'name': Destroy_Task, 'duration_secs': 0.337868} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.005741] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Destroyed the VM [ 997.005986] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Deleting Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 997.006268] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-255007a8-6ed3-4d4c-ac7e-002813450a68 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.013772] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 997.013772] env[63293]: value = "task-1328069" [ 997.013772] env[63293]: _type = "Task" [ 997.013772] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.022543] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328069, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.108746] env[63293]: DEBUG nova.compute.manager [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 997.123194] env[63293]: DEBUG nova.network.neutron [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Successfully created port: 69771278-cd61-43d4-b53c-76544685f282 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.305034] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "refresh_cache-bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.305196] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "refresh_cache-bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.305452] env[63293]: DEBUG nova.network.neutron [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 997.400266] env[63293]: DEBUG oslo_vmware.api [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328068, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.404864] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "d753598d-a92a-4515-9ad1-d386294f7a99" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.405594] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.405594] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "d753598d-a92a-4515-9ad1-d386294f7a99-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.405594] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.405855] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.408113] env[63293]: INFO nova.compute.manager [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Terminating instance [ 997.409815] env[63293]: DEBUG nova.scheduler.client.report [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 997.416020] env[63293]: DEBUG nova.compute.manager [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 997.416020] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.416020] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df0d37a-729c-496f-b6e0-7f129922a3a6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.424940] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.425255] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1826f6fa-c98b-4d49-9291-fc4804724ba0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.433496] env[63293]: DEBUG oslo_vmware.api [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 997.433496] env[63293]: value = "task-1328070" [ 997.433496] env[63293]: _type = "Task" [ 997.433496] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.444898] env[63293]: DEBUG oslo_vmware.api [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328070, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.525289] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328069, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.848918] env[63293]: DEBUG nova.network.neutron [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 997.901759] env[63293]: DEBUG oslo_vmware.api [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328068, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.918163] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.813s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.939111] env[63293]: INFO nova.scheduler.client.report [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleted allocations for instance ad585ebb-2072-45df-b645-94c9fa93576b [ 997.952304] env[63293]: DEBUG oslo_vmware.api [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328070, 'name': PowerOffVM_Task, 'duration_secs': 0.342743} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.952822] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 997.953061] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 997.953354] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-448bd00c-af21-4cc4-94c2-04908cd45ff3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.972640] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquiring lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.972943] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.027419] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328069, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.042561] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.042950] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.043245] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleting the datastore file [datastore1] d753598d-a92a-4515-9ad1-d386294f7a99 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.044958] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55a29363-8bb9-412d-9ecf-9da001a4eae1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.054539] env[63293]: DEBUG oslo_vmware.api [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 998.054539] env[63293]: value = "task-1328072" [ 998.054539] env[63293]: _type = "Task" [ 998.054539] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.064590] env[63293]: DEBUG oslo_vmware.api [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328072, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.112999] env[63293]: DEBUG nova.network.neutron [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Updating instance_info_cache with network_info: [{"id": "12ae71a2-a302-4223-8d65-630cde456181", "address": "fa:16:3e:88:c0:8b", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ae71a2-a3", "ovs_interfaceid": "12ae71a2-a302-4223-8d65-630cde456181", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.119117] env[63293]: DEBUG nova.compute.manager [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 998.151436] env[63293]: DEBUG nova.virt.hardware [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 998.151747] env[63293]: DEBUG nova.virt.hardware [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 998.151912] env[63293]: DEBUG nova.virt.hardware [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.152139] env[63293]: DEBUG nova.virt.hardware [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 998.152267] env[63293]: DEBUG nova.virt.hardware [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.152514] env[63293]: DEBUG nova.virt.hardware [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 998.152758] env[63293]: DEBUG nova.virt.hardware [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 998.152928] env[63293]: DEBUG nova.virt.hardware [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 998.153115] env[63293]: DEBUG nova.virt.hardware [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 998.153288] env[63293]: DEBUG nova.virt.hardware [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 998.153466] env[63293]: DEBUG nova.virt.hardware [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 998.154374] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bccf54-a00d-49f1-9762-304e2c648ad5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.171277] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36a644d-54af-4b74-9bea-365ffb134d28 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.400761] env[63293]: DEBUG oslo_vmware.api [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328068, 'name': ReconfigVM_Task, 'duration_secs': 1.101218} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.401349] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.401589] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Reconfigured VM to attach interface {{(pid=63293) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 998.456354] env[63293]: DEBUG oslo_concurrency.lockutils [None req-21a52d53-da4c-4673-ad2c-41ef540378fa tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "ad585ebb-2072-45df-b645-94c9fa93576b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.399s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.474923] env[63293]: DEBUG nova.compute.manager [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 998.531744] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328069, 'name': RemoveSnapshot_Task, 'duration_secs': 1.168098} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.532242] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Deleted Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 998.532702] env[63293]: DEBUG nova.compute.manager [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.534060] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2721c4b8-48a9-4d4a-af03-dda43271b61d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.568890] env[63293]: DEBUG oslo_vmware.api [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328072, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.271997} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.569349] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 998.569701] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 998.569984] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 998.570372] env[63293]: INFO nova.compute.manager [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Took 1.16 seconds to destroy the instance on the hypervisor. [ 998.570728] env[63293]: DEBUG oslo.service.loopingcall [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.571045] env[63293]: DEBUG nova.compute.manager [-] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 998.571214] env[63293]: DEBUG nova.network.neutron [-] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 998.616323] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "refresh_cache-bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.616922] env[63293]: DEBUG nova.compute.manager [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Instance network_info: |[{"id": "12ae71a2-a302-4223-8d65-630cde456181", "address": "fa:16:3e:88:c0:8b", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ae71a2-a3", "ovs_interfaceid": "12ae71a2-a302-4223-8d65-630cde456181", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 998.617575] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:c0:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ef746c57-cd18-4883-a0e9-c52937aaf41d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '12ae71a2-a302-4223-8d65-630cde456181', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 998.627309] env[63293]: DEBUG oslo.service.loopingcall [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.628244] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 998.628581] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ce0a794-2429-4e11-8d69-bd0239a29088 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.653896] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 998.653896] env[63293]: value = "task-1328073" [ 998.653896] env[63293]: _type = "Task" [ 998.653896] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.663801] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328073, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.668360] env[63293]: DEBUG nova.compute.manager [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Stashing vm_state: active {{(pid=63293) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 998.907224] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e0bc9e4e-26a4-4f20-99e7-8135abd53656 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-f9ccb216-1435-44c5-ab34-a6388d794551-c2fa0388-8ddf-497f-8200-903c0fde5a7b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.662s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.999989] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.000335] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.007129] env[63293]: INFO nova.compute.claims [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 999.053903] env[63293]: INFO nova.compute.manager [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Shelve offloading [ 999.055789] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 999.055956] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70dcb422-0e38-45e9-8123-3595536eaf59 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.067878] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 999.067878] env[63293]: value = "task-1328074" [ 999.067878] env[63293]: _type = "Task" [ 999.067878] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.096052] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] VM already powered off {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 999.096379] env[63293]: DEBUG nova.compute.manager [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.097294] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476af767-4bd1-42b5-bb61-17e02f8be0f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.106709] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.106903] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.107096] env[63293]: DEBUG nova.network.neutron [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 999.122779] env[63293]: DEBUG nova.network.neutron [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Successfully updated port: 69771278-cd61-43d4-b53c-76544685f282 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.127383] env[63293]: DEBUG nova.compute.manager [req-5fb70382-509b-4416-b8e5-9979b6a2a0ec req-9dcf917f-61d3-4b43-bdd2-780b5206593c service nova] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Received event network-changed-12ae71a2-a302-4223-8d65-630cde456181 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.127383] env[63293]: DEBUG nova.compute.manager [req-5fb70382-509b-4416-b8e5-9979b6a2a0ec req-9dcf917f-61d3-4b43-bdd2-780b5206593c service nova] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Refreshing instance network info cache due to event network-changed-12ae71a2-a302-4223-8d65-630cde456181. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 999.127383] env[63293]: DEBUG oslo_concurrency.lockutils [req-5fb70382-509b-4416-b8e5-9979b6a2a0ec req-9dcf917f-61d3-4b43-bdd2-780b5206593c service nova] Acquiring lock "refresh_cache-bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.127383] env[63293]: DEBUG oslo_concurrency.lockutils [req-5fb70382-509b-4416-b8e5-9979b6a2a0ec req-9dcf917f-61d3-4b43-bdd2-780b5206593c service nova] Acquired lock "refresh_cache-bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.127383] env[63293]: DEBUG nova.network.neutron [req-5fb70382-509b-4416-b8e5-9979b6a2a0ec req-9dcf917f-61d3-4b43-bdd2-780b5206593c service nova] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Refreshing network info cache for port 12ae71a2-a302-4223-8d65-630cde456181 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 999.171754] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328073, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.192215] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.565414] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.565414] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.629031] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "refresh_cache-71154ca4-ea2c-4187-8319-0bfb9b289eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.629031] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "refresh_cache-71154ca4-ea2c-4187-8319-0bfb9b289eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.629031] env[63293]: DEBUG nova.network.neutron [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 999.665444] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328073, 'name': CreateVM_Task, 'duration_secs': 0.680738} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.665608] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 999.666396] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.666532] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.666816] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 999.670023] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5950d5b5-e982-4ee4-9bfb-8407410dac6b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.673169] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 999.673169] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527d3840-ce1d-8a4f-4da1-2bd7138b11b1" [ 999.673169] env[63293]: _type = "Task" [ 999.673169] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.683809] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527d3840-ce1d-8a4f-4da1-2bd7138b11b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.945648] env[63293]: DEBUG nova.network.neutron [-] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.067707] env[63293]: DEBUG nova.compute.manager [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1000.117503] env[63293]: DEBUG nova.network.neutron [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updating instance_info_cache with network_info: [{"id": "4c817793-6a1c-4d84-ac42-0487913a257b", "address": "fa:16:3e:18:84:4b", "network": {"id": "8af42ccf-7ab8-424b-a290-8383c292ef85", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1850566934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eec85c5992d64d62a90e290a1aa2d441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c817793-6a", "ovs_interfaceid": "4c817793-6a1c-4d84-ac42-0487913a257b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.122115] env[63293]: DEBUG nova.network.neutron [req-5fb70382-509b-4416-b8e5-9979b6a2a0ec req-9dcf917f-61d3-4b43-bdd2-780b5206593c service nova] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Updated VIF entry in instance network info cache for port 12ae71a2-a302-4223-8d65-630cde456181. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1000.122730] env[63293]: DEBUG nova.network.neutron [req-5fb70382-509b-4416-b8e5-9979b6a2a0ec req-9dcf917f-61d3-4b43-bdd2-780b5206593c service nova] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Updating instance_info_cache with network_info: [{"id": "12ae71a2-a302-4223-8d65-630cde456181", "address": "fa:16:3e:88:c0:8b", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ae71a2-a3", "ovs_interfaceid": "12ae71a2-a302-4223-8d65-630cde456181", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.184765] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527d3840-ce1d-8a4f-4da1-2bd7138b11b1, 'name': SearchDatastore_Task, 'duration_secs': 0.016819} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.184765] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.184988] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.185200] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.185398] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.185628] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.187186] env[63293]: DEBUG nova.network.neutron [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1000.192310] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95bac563-a78e-48ec-b80c-9d2adde2deaf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.203266] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.203482] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1000.204582] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d47aa5a-bbdf-4873-b309-3f842874256d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.216478] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1000.216478] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ace3dd-d513-0f4b-9260-2f862b106e69" [ 1000.216478] env[63293]: _type = "Task" [ 1000.216478] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.227237] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ace3dd-d513-0f4b-9260-2f862b106e69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.238313] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3029185f-e826-4b82-a9f9-1e33303e5d6c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.251804] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db783e9-e9f0-47db-b987-ae7f3f7ab27a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.286246] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d325f5-7b28-46ad-a715-82cc34b7b595 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.295823] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bdf4a1c-b691-4004-9fa6-9cfe530d6b08 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.310253] env[63293]: DEBUG nova.compute.provider_tree [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.362190] env[63293]: DEBUG nova.network.neutron [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Updating instance_info_cache with network_info: [{"id": "69771278-cd61-43d4-b53c-76544685f282", "address": "fa:16:3e:53:36:83", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69771278-cd", "ovs_interfaceid": "69771278-cd61-43d4-b53c-76544685f282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.451890] env[63293]: INFO nova.compute.manager [-] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Took 1.88 seconds to deallocate network for instance. [ 1000.530923] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "interface-f9ccb216-1435-44c5-ab34-a6388d794551-c2fa0388-8ddf-497f-8200-903c0fde5a7b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.531251] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-f9ccb216-1435-44c5-ab34-a6388d794551-c2fa0388-8ddf-497f-8200-903c0fde5a7b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.586668] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.624726] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.626791] env[63293]: DEBUG oslo_concurrency.lockutils [req-5fb70382-509b-4416-b8e5-9979b6a2a0ec req-9dcf917f-61d3-4b43-bdd2-780b5206593c service nova] Releasing lock "refresh_cache-bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.729015] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ace3dd-d513-0f4b-9260-2f862b106e69, 'name': SearchDatastore_Task, 'duration_secs': 0.010877} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.730039] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aebadfc3-4292-4331-8a01-4533823beaf2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.737622] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1000.737622] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e4af50-a6b9-19fb-fbdb-50b8cf5fdf1d" [ 1000.737622] env[63293]: _type = "Task" [ 1000.737622] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.747282] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e4af50-a6b9-19fb-fbdb-50b8cf5fdf1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.815185] env[63293]: DEBUG nova.scheduler.client.report [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.865275] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "refresh_cache-71154ca4-ea2c-4187-8319-0bfb9b289eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.865619] env[63293]: DEBUG nova.compute.manager [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Instance network_info: |[{"id": "69771278-cd61-43d4-b53c-76544685f282", "address": "fa:16:3e:53:36:83", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69771278-cd", "ovs_interfaceid": "69771278-cd61-43d4-b53c-76544685f282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1000.866376] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:36:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aec0089a-ff85-4bef-bad8-c84de39af71a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69771278-cd61-43d4-b53c-76544685f282', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1000.874656] env[63293]: DEBUG oslo.service.loopingcall [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.874895] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1000.875141] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f6faf29-f724-4456-b5d9-c3c166418bda {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.898288] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1000.898288] env[63293]: value = "task-1328075" [ 1000.898288] env[63293]: _type = "Task" [ 1000.898288] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.906640] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328075, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.928370] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.929283] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c76f29-e151-465c-bb45-ad568ed237d9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.939764] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.939764] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7e1cc56-9214-4a84-982b-385c73a5f8b3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.959090] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.019842] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.020258] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.020494] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Deleting the datastore file [datastore2] f429869f-9e9f-443e-a980-67a80c3bd799 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.020795] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d1bd907-acab-4710-a2b1-884767a5335c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.030259] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1001.030259] env[63293]: value = "task-1328077" [ 1001.030259] env[63293]: _type = "Task" [ 1001.030259] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.034376] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.034559] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.038243] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d0ba31-6676-4dc9-9e85-8e328f8c094c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.041152] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328077, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.056906] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c20fa3-af74-44bb-9c4c-121a799dc65d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.090745] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Reconfiguring VM to detach interface {{(pid=63293) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1001.091086] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-902e5b42-1d36-4586-9968-dfc5dd86bb8c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.112226] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 1001.112226] env[63293]: value = "task-1328078" [ 1001.112226] env[63293]: _type = "Task" [ 1001.112226] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.121571] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.162109] env[63293]: DEBUG nova.compute.manager [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Received event network-vif-plugged-69771278-cd61-43d4-b53c-76544685f282 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.162382] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Acquiring lock "71154ca4-ea2c-4187-8319-0bfb9b289eab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.162648] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Lock "71154ca4-ea2c-4187-8319-0bfb9b289eab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.162867] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Lock "71154ca4-ea2c-4187-8319-0bfb9b289eab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.163093] env[63293]: DEBUG nova.compute.manager [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] No waiting events found dispatching network-vif-plugged-69771278-cd61-43d4-b53c-76544685f282 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1001.163321] env[63293]: WARNING nova.compute.manager [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Received unexpected event network-vif-plugged-69771278-cd61-43d4-b53c-76544685f282 for instance with vm_state building and task_state spawning. [ 1001.163546] env[63293]: DEBUG nova.compute.manager [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Received event network-changed-69771278-cd61-43d4-b53c-76544685f282 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.163764] env[63293]: DEBUG nova.compute.manager [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Refreshing instance network info cache due to event network-changed-69771278-cd61-43d4-b53c-76544685f282. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1001.164016] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Acquiring lock "refresh_cache-71154ca4-ea2c-4187-8319-0bfb9b289eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.164205] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Acquired lock "refresh_cache-71154ca4-ea2c-4187-8319-0bfb9b289eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.164720] env[63293]: DEBUG nova.network.neutron [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Refreshing network info cache for port 69771278-cd61-43d4-b53c-76544685f282 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1001.250246] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e4af50-a6b9-19fb-fbdb-50b8cf5fdf1d, 'name': SearchDatastore_Task, 'duration_secs': 0.010979} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.250632] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.254048] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7/bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1001.254048] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f3b33844-6ad6-4886-b678-aced8798507c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.259095] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1001.259095] env[63293]: value = "task-1328079" [ 1001.259095] env[63293]: _type = "Task" [ 1001.259095] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.267600] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328079, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.321098] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.321509] env[63293]: DEBUG nova.compute.manager [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1001.324311] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.132s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.408671] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328075, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.539944] env[63293]: DEBUG oslo_vmware.api [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328077, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177184} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.540382] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.540617] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.540807] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.556928] env[63293]: INFO nova.scheduler.client.report [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Deleted allocations for instance f429869f-9e9f-443e-a980-67a80c3bd799 [ 1001.623540] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.771762] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328079, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.827984] env[63293]: DEBUG nova.compute.utils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1001.830158] env[63293]: DEBUG nova.compute.manager [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1001.830158] env[63293]: DEBUG nova.network.neutron [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1001.834189] env[63293]: INFO nova.compute.claims [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.907285] env[63293]: DEBUG nova.policy [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a54f50e132ab44ceac9ce5a58a751873', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97cb1a48336448ad866206ab157778c2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 1001.915316] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328075, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.985791] env[63293]: DEBUG nova.network.neutron [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Updated VIF entry in instance network info cache for port 69771278-cd61-43d4-b53c-76544685f282. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1001.986202] env[63293]: DEBUG nova.network.neutron [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Updating instance_info_cache with network_info: [{"id": "69771278-cd61-43d4-b53c-76544685f282", "address": "fa:16:3e:53:36:83", "network": {"id": "c96ca127-6af7-4501-93a1-1efea91cafd0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1751494559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fac34e49d5dc49e7a7055c998d5b6766", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aec0089a-ff85-4bef-bad8-c84de39af71a", "external-id": "nsx-vlan-transportzone-758", "segmentation_id": 758, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69771278-cd", "ovs_interfaceid": "69771278-cd61-43d4-b53c-76544685f282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.015729] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.015978] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.062802] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.124704] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.274319] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328079, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.339386] env[63293]: DEBUG nova.compute.manager [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1002.343603] env[63293]: INFO nova.compute.resource_tracker [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating resource usage from migration 42d73401-baaf-4819-bf86-bb2fc639388e [ 1002.415253] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328075, 'name': CreateVM_Task, 'duration_secs': 1.339707} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.415544] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1002.416297] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.416550] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.416933] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1002.417268] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54695026-fecc-41ae-8b91-9557bd02ecae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.425149] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 1002.425149] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52beec4d-8b93-205a-1b23-5c6a9dac7800" [ 1002.425149] env[63293]: _type = "Task" [ 1002.425149] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.433391] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52beec4d-8b93-205a-1b23-5c6a9dac7800, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.492310] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Releasing lock "refresh_cache-71154ca4-ea2c-4187-8319-0bfb9b289eab" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.492581] env[63293]: DEBUG nova.compute.manager [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Received event network-vif-deleted-d41cf464-2396-4e22-80d1-3c1602939528 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.492775] env[63293]: DEBUG nova.compute.manager [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Received event network-vif-unplugged-4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.492963] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Acquiring lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.493390] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.493612] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.493968] env[63293]: DEBUG nova.compute.manager [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] No waiting events found dispatching network-vif-unplugged-4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1002.494156] env[63293]: WARNING nova.compute.manager [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Received unexpected event network-vif-unplugged-4c817793-6a1c-4d84-ac42-0487913a257b for instance with vm_state shelved and task_state shelving_offloading. [ 1002.494436] env[63293]: DEBUG nova.compute.manager [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Received event network-changed-4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.494679] env[63293]: DEBUG nova.compute.manager [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Refreshing instance network info cache due to event network-changed-4c817793-6a1c-4d84-ac42-0487913a257b. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1002.494901] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Acquiring lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.495089] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Acquired lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.495293] env[63293]: DEBUG nova.network.neutron [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Refreshing network info cache for port 4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1002.507468] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a4a81b-9174-44a2-971f-9a097bb01279 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.516119] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fc4e376-b892-4f89-9160-a4c18b19957b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.519312] env[63293]: DEBUG nova.compute.manager [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1002.558092] env[63293]: DEBUG nova.network.neutron [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Successfully created port: 8147f046-74b1-4a9e-b437-02f3a5ee8fe3 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1002.560618] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-598e3ead-a4d2-4afe-a24e-9c386c32dce6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.571323] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f09828-508a-448c-8159-3847bcb44eb3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.588543] env[63293]: DEBUG nova.compute.provider_tree [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.626059] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.772679] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328079, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.939087] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52beec4d-8b93-205a-1b23-5c6a9dac7800, 'name': SearchDatastore_Task, 'duration_secs': 0.42984} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.940363] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.940363] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1002.940363] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.940363] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.940515] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1002.940822] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07ec89a1-76cc-400d-96c8-87b73fa47880 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.958032] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1002.958032] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1002.958328] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf58a4fb-0486-4cd1-bbed-269bd64dedef {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.965687] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 1002.965687] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267e268-e9be-b544-09d7-8dda9e2a35cf" [ 1002.965687] env[63293]: _type = "Task" [ 1002.965687] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.976334] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267e268-e9be-b544-09d7-8dda9e2a35cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.042431] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.091843] env[63293]: DEBUG nova.scheduler.client.report [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.125059] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.253306] env[63293]: DEBUG nova.network.neutron [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updated VIF entry in instance network info cache for port 4c817793-6a1c-4d84-ac42-0487913a257b. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1003.253702] env[63293]: DEBUG nova.network.neutron [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updating instance_info_cache with network_info: [{"id": "4c817793-6a1c-4d84-ac42-0487913a257b", "address": "fa:16:3e:18:84:4b", "network": {"id": "8af42ccf-7ab8-424b-a290-8383c292ef85", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1850566934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eec85c5992d64d62a90e290a1aa2d441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4c817793-6a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.273354] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328079, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.779359} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.273613] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7/bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1003.273833] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.274106] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-554558dc-1da9-4444-928d-2d3dd686a84a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.283797] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1003.283797] env[63293]: value = "task-1328080" [ 1003.283797] env[63293]: _type = "Task" [ 1003.283797] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.294439] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328080, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.339128] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "f429869f-9e9f-443e-a980-67a80c3bd799" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.353706] env[63293]: DEBUG nova.compute.manager [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1003.379949] env[63293]: DEBUG nova.virt.hardware [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1003.380261] env[63293]: DEBUG nova.virt.hardware [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1003.380428] env[63293]: DEBUG nova.virt.hardware [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1003.380639] env[63293]: DEBUG nova.virt.hardware [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1003.380791] env[63293]: DEBUG nova.virt.hardware [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1003.380943] env[63293]: DEBUG nova.virt.hardware [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1003.381166] env[63293]: DEBUG nova.virt.hardware [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1003.381362] env[63293]: DEBUG nova.virt.hardware [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1003.381554] env[63293]: DEBUG nova.virt.hardware [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1003.381725] env[63293]: DEBUG nova.virt.hardware [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1003.381899] env[63293]: DEBUG nova.virt.hardware [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1003.382797] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afa46f4-3b9a-4342-9357-c37ea8ab3312 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.391736] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ead290-71a8-4abc-8c78-dc90c3fda3db {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.477839] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5267e268-e9be-b544-09d7-8dda9e2a35cf, 'name': SearchDatastore_Task, 'duration_secs': 0.050397} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.478084] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dcd9582-d55e-45f1-8997-ba4a51add48a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.484769] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 1003.484769] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523abc41-235a-d207-d164-0b2b3bbb986a" [ 1003.484769] env[63293]: _type = "Task" [ 1003.484769] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.496251] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523abc41-235a-d207-d164-0b2b3bbb986a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.596992] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.272s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.597329] env[63293]: INFO nova.compute.manager [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Migrating [ 1003.604533] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.018s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.606654] env[63293]: INFO nova.compute.claims [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1003.628063] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.756888] env[63293]: DEBUG oslo_concurrency.lockutils [req-2e39fc2a-0ca2-4d80-9801-ec0893c90558 req-98490707-2d02-4686-ad4c-465d52602c87 service nova] Releasing lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.793908] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328080, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.971837] env[63293]: DEBUG nova.compute.manager [req-cf8d2696-348a-45ac-ba8d-9dc09d56eae0 req-291ab017-15d6-4720-bed5-f3b9dae48360 service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Received event network-vif-plugged-8147f046-74b1-4a9e-b437-02f3a5ee8fe3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.972104] env[63293]: DEBUG oslo_concurrency.lockutils [req-cf8d2696-348a-45ac-ba8d-9dc09d56eae0 req-291ab017-15d6-4720-bed5-f3b9dae48360 service nova] Acquiring lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.972378] env[63293]: DEBUG oslo_concurrency.lockutils [req-cf8d2696-348a-45ac-ba8d-9dc09d56eae0 req-291ab017-15d6-4720-bed5-f3b9dae48360 service nova] Lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.972496] env[63293]: DEBUG oslo_concurrency.lockutils [req-cf8d2696-348a-45ac-ba8d-9dc09d56eae0 req-291ab017-15d6-4720-bed5-f3b9dae48360 service nova] Lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.972673] env[63293]: DEBUG nova.compute.manager [req-cf8d2696-348a-45ac-ba8d-9dc09d56eae0 req-291ab017-15d6-4720-bed5-f3b9dae48360 service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] No waiting events found dispatching network-vif-plugged-8147f046-74b1-4a9e-b437-02f3a5ee8fe3 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1003.972843] env[63293]: WARNING nova.compute.manager [req-cf8d2696-348a-45ac-ba8d-9dc09d56eae0 req-291ab017-15d6-4720-bed5-f3b9dae48360 service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Received unexpected event network-vif-plugged-8147f046-74b1-4a9e-b437-02f3a5ee8fe3 for instance with vm_state building and task_state spawning. [ 1003.996331] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523abc41-235a-d207-d164-0b2b3bbb986a, 'name': SearchDatastore_Task, 'duration_secs': 0.377614} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.996598] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.996854] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 71154ca4-ea2c-4187-8319-0bfb9b289eab/71154ca4-ea2c-4187-8319-0bfb9b289eab.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1003.997129] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a2c6445-fd15-413d-b1ae-8806c4bfbb29 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.004593] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 1004.004593] env[63293]: value = "task-1328081" [ 1004.004593] env[63293]: _type = "Task" [ 1004.004593] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.013420] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328081, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.065009] env[63293]: DEBUG nova.network.neutron [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Successfully updated port: 8147f046-74b1-4a9e-b437-02f3a5ee8fe3 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1004.121634] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.121634] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.121634] env[63293]: DEBUG nova.network.neutron [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1004.132719] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.296127] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328080, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.585561} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.296468] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.297758] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f26f7b98-358c-4949-9cd0-8b07acda58a6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.321740] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7/bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.322116] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ba8c72d-8c47-46c1-b951-b51e980efd9e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.343846] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1004.343846] env[63293]: value = "task-1328082" [ 1004.343846] env[63293]: _type = "Task" [ 1004.343846] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.356635] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328082, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.518962] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328081, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.568890] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquiring lock "refresh_cache-2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.569111] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquired lock "refresh_cache-2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.569361] env[63293]: DEBUG nova.network.neutron [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1004.637792] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.794456] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16f0833-2bed-4cb9-9819-77f3b4fa9d24 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.802911] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a66262-42c4-49e9-a1f0-3b7e7339c84b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.835390] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab35034d-9ba6-4a85-a58b-e27e9322bca9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.844184] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a539e7-9a95-47ba-b0f0-b4e477194e30 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.856320] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328082, 'name': ReconfigVM_Task, 'duration_secs': 0.384017} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.864503] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Reconfigured VM instance instance-00000061 to attach disk [datastore1] bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7/bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.865311] env[63293]: DEBUG nova.compute.provider_tree [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.866891] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9545597d-e56f-4ac7-9b1e-e573f47c45f7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.877079] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1004.877079] env[63293]: value = "task-1328083" [ 1004.877079] env[63293]: _type = "Task" [ 1004.877079] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.887176] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328083, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.913930] env[63293]: DEBUG nova.network.neutron [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance_info_cache with network_info: [{"id": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "address": "fa:16:3e:c2:85:7d", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83a30fb0-a4", "ovs_interfaceid": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.016885] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328081, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.629078} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.017127] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 71154ca4-ea2c-4187-8319-0bfb9b289eab/71154ca4-ea2c-4187-8319-0bfb9b289eab.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1005.017332] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.017591] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21c49b89-8a04-4230-8f3a-d175d2654715 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.026297] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 1005.026297] env[63293]: value = "task-1328084" [ 1005.026297] env[63293]: _type = "Task" [ 1005.026297] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.035045] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328084, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.105584] env[63293]: DEBUG nova.network.neutron [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1005.133624] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.239134] env[63293]: DEBUG nova.network.neutron [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Updating instance_info_cache with network_info: [{"id": "8147f046-74b1-4a9e-b437-02f3a5ee8fe3", "address": "fa:16:3e:c2:a4:a0", "network": {"id": "39a6f586-5383-45ed-999b-db33c8f4d9f3", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-884746607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97cb1a48336448ad866206ab157778c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8147f046-74", "ovs_interfaceid": "8147f046-74b1-4a9e-b437-02f3a5ee8fe3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.370973] env[63293]: DEBUG nova.scheduler.client.report [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.388658] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328083, 'name': Rename_Task, 'duration_secs': 0.165876} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.389665] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1005.390166] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18df9332-dc76-4f55-8fa3-6de076f25910 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.401051] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1005.401051] env[63293]: value = "task-1328085" [ 1005.401051] env[63293]: _type = "Task" [ 1005.401051] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.410299] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328085, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.417047] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.537116] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328084, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089164} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.537429] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1005.538274] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d80f4c9-4cf9-4579-806d-b4e84da54573 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.561486] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 71154ca4-ea2c-4187-8319-0bfb9b289eab/71154ca4-ea2c-4187-8319-0bfb9b289eab.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.562201] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6225b65b-ff66-4d30-bc6d-3131c6673e19 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.584229] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 1005.584229] env[63293]: value = "task-1328086" [ 1005.584229] env[63293]: _type = "Task" [ 1005.584229] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.593117] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328086, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.634868] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.742370] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Releasing lock "refresh_cache-2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.742736] env[63293]: DEBUG nova.compute.manager [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Instance network_info: |[{"id": "8147f046-74b1-4a9e-b437-02f3a5ee8fe3", "address": "fa:16:3e:c2:a4:a0", "network": {"id": "39a6f586-5383-45ed-999b-db33c8f4d9f3", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-884746607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97cb1a48336448ad866206ab157778c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8147f046-74", "ovs_interfaceid": "8147f046-74b1-4a9e-b437-02f3a5ee8fe3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1005.743242] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:a4:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8147f046-74b1-4a9e-b437-02f3a5ee8fe3', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1005.752492] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Creating folder: Project (97cb1a48336448ad866206ab157778c2). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1005.752837] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-820ed56c-e172-47e9-822d-b715a666a534 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.767837] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Created folder: Project (97cb1a48336448ad866206ab157778c2) in parent group-v283678. [ 1005.768067] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Creating folder: Instances. Parent ref: group-v283825. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1005.768334] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6733a65e-a8dc-4d4b-8eda-10099f804e79 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.779776] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Created folder: Instances in parent group-v283825. [ 1005.780053] env[63293]: DEBUG oslo.service.loopingcall [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.780356] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1005.780574] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56b7f3cb-75a4-4ed8-b6e1-3dd491a8aa5f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.800913] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1005.800913] env[63293]: value = "task-1328089" [ 1005.800913] env[63293]: _type = "Task" [ 1005.800913] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.809138] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328089, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.876548] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.272s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.877151] env[63293]: DEBUG nova.compute.manager [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1005.880907] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.922s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.881303] env[63293]: DEBUG nova.objects.instance [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lazy-loading 'resources' on Instance uuid d753598d-a92a-4515-9ad1-d386294f7a99 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.923638] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328085, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.998331] env[63293]: DEBUG nova.compute.manager [req-f3f45b40-e88f-485c-b3d1-9a4655ab9082 req-681039f5-3fca-4d36-bce8-f1191d81314a service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Received event network-changed-8147f046-74b1-4a9e-b437-02f3a5ee8fe3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.998451] env[63293]: DEBUG nova.compute.manager [req-f3f45b40-e88f-485c-b3d1-9a4655ab9082 req-681039f5-3fca-4d36-bce8-f1191d81314a service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Refreshing instance network info cache due to event network-changed-8147f046-74b1-4a9e-b437-02f3a5ee8fe3. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1005.998680] env[63293]: DEBUG oslo_concurrency.lockutils [req-f3f45b40-e88f-485c-b3d1-9a4655ab9082 req-681039f5-3fca-4d36-bce8-f1191d81314a service nova] Acquiring lock "refresh_cache-2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.998833] env[63293]: DEBUG oslo_concurrency.lockutils [req-f3f45b40-e88f-485c-b3d1-9a4655ab9082 req-681039f5-3fca-4d36-bce8-f1191d81314a service nova] Acquired lock "refresh_cache-2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.999012] env[63293]: DEBUG nova.network.neutron [req-f3f45b40-e88f-485c-b3d1-9a4655ab9082 req-681039f5-3fca-4d36-bce8-f1191d81314a service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Refreshing network info cache for port 8147f046-74b1-4a9e-b437-02f3a5ee8fe3 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1006.095943] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328086, 'name': ReconfigVM_Task, 'duration_secs': 0.376817} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.096199] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 71154ca4-ea2c-4187-8319-0bfb9b289eab/71154ca4-ea2c-4187-8319-0bfb9b289eab.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.096866] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-56ceb962-72bd-4ae6-9933-58e86508e441 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.103183] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 1006.103183] env[63293]: value = "task-1328090" [ 1006.103183] env[63293]: _type = "Task" [ 1006.103183] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.111960] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328090, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.135801] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.312112] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328089, 'name': CreateVM_Task, 'duration_secs': 0.363392} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.312344] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1006.313165] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.313339] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.313674] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1006.313930] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3e0c17f-2473-40c7-8ccd-22989a81ad19 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.318789] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for the task: (returnval){ [ 1006.318789] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c0d01d-2f7c-a784-6d0a-7739e4a117b3" [ 1006.318789] env[63293]: _type = "Task" [ 1006.318789] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.327101] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c0d01d-2f7c-a784-6d0a-7739e4a117b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.385028] env[63293]: DEBUG nova.compute.utils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1006.386473] env[63293]: DEBUG nova.compute.manager [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1006.386643] env[63293]: DEBUG nova.network.neutron [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1006.416021] env[63293]: DEBUG oslo_vmware.api [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328085, 'name': PowerOnVM_Task, 'duration_secs': 0.549268} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.416021] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1006.416226] env[63293]: INFO nova.compute.manager [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Took 10.65 seconds to spawn the instance on the hypervisor. [ 1006.416908] env[63293]: DEBUG nova.compute.manager [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.417335] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409292df-1f2a-4661-9c72-c46914a6e927 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.441326] env[63293]: DEBUG nova.policy [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ebf82e99427d4171a4d510f7d3c966d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1687c33183e74b6ba70e021b6879cb93', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 1006.554421] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b2fe95-8f23-4f42-9932-ae68050fbbc3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.566289] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07bedf4-55ef-4493-a3aa-32babdf488b9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.599956] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc78def1-4153-424a-bd66-7390be62958f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.610655] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7bfe9d-c1a6-465e-b9e0-4f4e59369ae7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.617516] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328090, 'name': Rename_Task, 'duration_secs': 0.345915} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.618199] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1006.618524] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9caf0774-adec-420f-a78e-b98ddbabc995 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.631116] env[63293]: DEBUG nova.compute.provider_tree [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.638780] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 1006.638780] env[63293]: value = "task-1328091" [ 1006.638780] env[63293]: _type = "Task" [ 1006.638780] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.653250] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.656756] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328091, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.832406] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c0d01d-2f7c-a784-6d0a-7739e4a117b3, 'name': SearchDatastore_Task, 'duration_secs': 0.00976} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.832406] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.832406] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1006.832635] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.832635] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.832850] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1006.833168] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-147d73cf-11f2-45d6-b083-9c99aaca72d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.843626] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1006.843888] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1006.844702] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4827af53-1ccd-4b1e-970f-00eb7744245b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.851496] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for the task: (returnval){ [ 1006.851496] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5240b140-c5cc-0b38-e23d-ba327988d31a" [ 1006.851496] env[63293]: _type = "Task" [ 1006.851496] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.860305] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5240b140-c5cc-0b38-e23d-ba327988d31a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.892602] env[63293]: DEBUG nova.compute.manager [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1006.933700] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b32da5ba-d60a-46fe-a4f5-2bc5e2181dd3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.963300] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance 'd605e066-806a-4aac-9af6-f79d542ab2e3' progress to 0 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1006.967289] env[63293]: DEBUG nova.network.neutron [req-f3f45b40-e88f-485c-b3d1-9a4655ab9082 req-681039f5-3fca-4d36-bce8-f1191d81314a service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Updated VIF entry in instance network info cache for port 8147f046-74b1-4a9e-b437-02f3a5ee8fe3. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1006.967627] env[63293]: DEBUG nova.network.neutron [req-f3f45b40-e88f-485c-b3d1-9a4655ab9082 req-681039f5-3fca-4d36-bce8-f1191d81314a service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Updating instance_info_cache with network_info: [{"id": "8147f046-74b1-4a9e-b437-02f3a5ee8fe3", "address": "fa:16:3e:c2:a4:a0", "network": {"id": "39a6f586-5383-45ed-999b-db33c8f4d9f3", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-884746607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97cb1a48336448ad866206ab157778c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8147f046-74", "ovs_interfaceid": "8147f046-74b1-4a9e-b437-02f3a5ee8fe3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.971915] env[63293]: INFO nova.compute.manager [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Took 18.42 seconds to build instance. [ 1006.998725] env[63293]: DEBUG nova.network.neutron [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Successfully created port: c589de77-25fc-43f7-9480-e6c97bbda93d {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1007.133379] env[63293]: DEBUG nova.scheduler.client.report [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.142667] env[63293]: DEBUG oslo_vmware.api [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328078, 'name': ReconfigVM_Task, 'duration_secs': 5.800577} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.145888] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.146138] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Reconfigured VM to detach interface {{(pid=63293) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1007.154745] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328091, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.363764] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5240b140-c5cc-0b38-e23d-ba327988d31a, 'name': SearchDatastore_Task, 'duration_secs': 0.010602} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.364791] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-094107b8-9f7b-4766-8885-5a4cb44c5d4c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.370510] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for the task: (returnval){ [ 1007.370510] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523551fa-65b6-0510-ca80-5f0588eaf87c" [ 1007.370510] env[63293]: _type = "Task" [ 1007.370510] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.382668] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523551fa-65b6-0510-ca80-5f0588eaf87c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.476020] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.476020] env[63293]: DEBUG oslo_concurrency.lockutils [req-f3f45b40-e88f-485c-b3d1-9a4655ab9082 req-681039f5-3fca-4d36-bce8-f1191d81314a service nova] Releasing lock "refresh_cache-2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.476020] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74435813-cf10-40d1-a7fd-8fedc3be5441 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.476918] env[63293]: DEBUG oslo_concurrency.lockutils [None req-188c7a9e-a5d4-425e-85bc-b65dceaeeef0 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.940s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.483736] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 1007.483736] env[63293]: value = "task-1328092" [ 1007.483736] env[63293]: _type = "Task" [ 1007.483736] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.494205] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328092, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.638499] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.758s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.641537] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.580s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.641757] env[63293]: DEBUG nova.objects.instance [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lazy-loading 'resources' on Instance uuid f429869f-9e9f-443e-a980-67a80c3bd799 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.653261] env[63293]: DEBUG oslo_vmware.api [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328091, 'name': PowerOnVM_Task, 'duration_secs': 0.635067} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.653491] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1007.653632] env[63293]: INFO nova.compute.manager [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Took 9.53 seconds to spawn the instance on the hypervisor. [ 1007.653839] env[63293]: DEBUG nova.compute.manager [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.655010] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3e23b9-6ce7-4408-b786-4cc236328aba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.668719] env[63293]: INFO nova.scheduler.client.report [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleted allocations for instance d753598d-a92a-4515-9ad1-d386294f7a99 [ 1007.881947] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.882232] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.882442] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.882636] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.882807] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.884989] env[63293]: INFO nova.compute.manager [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Terminating instance [ 1007.886389] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523551fa-65b6-0510-ca80-5f0588eaf87c, 'name': SearchDatastore_Task, 'duration_secs': 0.018065} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.886682] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.886912] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50/2bf4a8a9-c46d-4512-96b7-b34b1bf69f50.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1007.887220] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf0a1612-3324-4a23-9f03-83e098a260fd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.889946] env[63293]: DEBUG nova.compute.manager [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1007.890125] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1007.890888] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe99cbca-75f4-4cc8-a277-f6ee205247f6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.899750] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.900969] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-189e2d86-2967-43b0-8a09-494d62e3930a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.902496] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for the task: (returnval){ [ 1007.902496] env[63293]: value = "task-1328093" [ 1007.902496] env[63293]: _type = "Task" [ 1007.902496] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.906811] env[63293]: DEBUG nova.compute.manager [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1007.912954] env[63293]: DEBUG oslo_vmware.api [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1007.912954] env[63293]: value = "task-1328094" [ 1007.912954] env[63293]: _type = "Task" [ 1007.912954] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.920329] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328093, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.927349] env[63293]: DEBUG oslo_vmware.api [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328094, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.942812] env[63293]: DEBUG nova.virt.hardware [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1007.943129] env[63293]: DEBUG nova.virt.hardware [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1007.943338] env[63293]: DEBUG nova.virt.hardware [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.943565] env[63293]: DEBUG nova.virt.hardware [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1007.943742] env[63293]: DEBUG nova.virt.hardware [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.943943] env[63293]: DEBUG nova.virt.hardware [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1007.944206] env[63293]: DEBUG nova.virt.hardware [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1007.944407] env[63293]: DEBUG nova.virt.hardware [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1007.944623] env[63293]: DEBUG nova.virt.hardware [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1007.944828] env[63293]: DEBUG nova.virt.hardware [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1007.945054] env[63293]: DEBUG nova.virt.hardware [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.946086] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8637822-fee9-429f-925e-c2a47990c31c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.955900] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7886f78-f272-4f60-88de-9f05268d3820 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.994860] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328092, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.144214] env[63293]: DEBUG nova.objects.instance [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lazy-loading 'numa_topology' on Instance uuid f429869f-9e9f-443e-a980-67a80c3bd799 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.176333] env[63293]: INFO nova.compute.manager [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Took 17.18 seconds to build instance. [ 1008.176572] env[63293]: DEBUG oslo_concurrency.lockutils [None req-be76b6b9-13ab-4a38-9409-b3af80a35c6b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "d753598d-a92a-4515-9ad1-d386294f7a99" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.771s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.413306] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328093, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465806} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.413732] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50/2bf4a8a9-c46d-4512-96b7-b34b1bf69f50.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1008.413908] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1008.417139] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8380b0c3-c5a1-4134-b34c-db3188702d8a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.424453] env[63293]: DEBUG oslo_vmware.api [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328094, 'name': PowerOffVM_Task, 'duration_secs': 0.201132} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.425711] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1008.425894] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1008.426204] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for the task: (returnval){ [ 1008.426204] env[63293]: value = "task-1328095" [ 1008.426204] env[63293]: _type = "Task" [ 1008.426204] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.426407] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2037e9a3-31b7-4f10-9e6a-6197783a1c0c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.436102] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328095, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.496023] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328092, 'name': PowerOffVM_Task, 'duration_secs': 0.644842} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.496339] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1008.496531] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance 'd605e066-806a-4aac-9af6-f79d542ab2e3' progress to 17 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1008.501154] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.501417] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.501634] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleting the datastore file [datastore1] bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.502140] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aeca0feb-59ea-4747-afb4-0c1d448e6061 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.509296] env[63293]: DEBUG oslo_vmware.api [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1008.509296] env[63293]: value = "task-1328097" [ 1008.509296] env[63293]: _type = "Task" [ 1008.509296] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.518073] env[63293]: DEBUG oslo_vmware.api [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328097, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.628395] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.628395] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquired lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.628395] env[63293]: DEBUG nova.network.neutron [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.646355] env[63293]: DEBUG nova.objects.base [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1008.678593] env[63293]: DEBUG oslo_concurrency.lockutils [None req-6fe90fe1-be79-4976-8b8f-48c2b819f922 tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "71154ca4-ea2c-4187-8319-0bfb9b289eab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.702s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.786478] env[63293]: DEBUG nova.compute.manager [req-21a1ec26-c691-4f5c-b70d-9bb93753794b req-f6b2d282-4c95-404a-a0b5-59caf1b2cd76 service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Received event network-vif-plugged-c589de77-25fc-43f7-9480-e6c97bbda93d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.786712] env[63293]: DEBUG oslo_concurrency.lockutils [req-21a1ec26-c691-4f5c-b70d-9bb93753794b req-f6b2d282-4c95-404a-a0b5-59caf1b2cd76 service nova] Acquiring lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.786919] env[63293]: DEBUG oslo_concurrency.lockutils [req-21a1ec26-c691-4f5c-b70d-9bb93753794b req-f6b2d282-4c95-404a-a0b5-59caf1b2cd76 service nova] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.787105] env[63293]: DEBUG oslo_concurrency.lockutils [req-21a1ec26-c691-4f5c-b70d-9bb93753794b req-f6b2d282-4c95-404a-a0b5-59caf1b2cd76 service nova] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.787280] env[63293]: DEBUG nova.compute.manager [req-21a1ec26-c691-4f5c-b70d-9bb93753794b req-f6b2d282-4c95-404a-a0b5-59caf1b2cd76 service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] No waiting events found dispatching network-vif-plugged-c589de77-25fc-43f7-9480-e6c97bbda93d {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1008.787448] env[63293]: WARNING nova.compute.manager [req-21a1ec26-c691-4f5c-b70d-9bb93753794b req-f6b2d282-4c95-404a-a0b5-59caf1b2cd76 service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Received unexpected event network-vif-plugged-c589de77-25fc-43f7-9480-e6c97bbda93d for instance with vm_state building and task_state spawning. [ 1008.795063] env[63293]: DEBUG nova.network.neutron [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Successfully updated port: c589de77-25fc-43f7-9480-e6c97bbda93d {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1008.812763] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab38a830-3599-4e11-8628-9a3632d2c972 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.821139] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82bffdd-dbce-4748-963d-dce7c0f3bff7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.855878] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c270603e-181d-49ec-9765-41891a2b8eed {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.870188] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f119a230-4e66-4c62-aa62-1a1390ea13ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.889182] env[63293]: DEBUG nova.compute.provider_tree [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.939204] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328095, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06333} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.939508] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1008.940402] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6637bb2-4bd3-4199-9612-7472fe869652 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.964587] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50/2bf4a8a9-c46d-4512-96b7-b34b1bf69f50.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1008.964895] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2e51853-f95a-4a0e-b191-9a04f4a4bad0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.986615] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for the task: (returnval){ [ 1008.986615] env[63293]: value = "task-1328098" [ 1008.986615] env[63293]: _type = "Task" [ 1008.986615] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.995644] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328098, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.004942] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:32:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1009.005213] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1009.005449] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.005713] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1009.005944] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.006270] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1009.006497] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1009.006765] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1009.007026] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1009.007213] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1009.007401] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.013554] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1a0b972-142d-4ec5-8862-1742bb467b30 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.033770] env[63293]: DEBUG oslo_vmware.api [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328097, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160112} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.035160] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.035367] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.035607] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.035800] env[63293]: INFO nova.compute.manager [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1009.036064] env[63293]: DEBUG oslo.service.loopingcall [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.036332] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 1009.036332] env[63293]: value = "task-1328099" [ 1009.036332] env[63293]: _type = "Task" [ 1009.036332] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.036620] env[63293]: DEBUG nova.compute.manager [-] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1009.036722] env[63293]: DEBUG nova.network.neutron [-] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1009.048668] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328099, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.082336] env[63293]: DEBUG oslo_concurrency.lockutils [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "f9ccb216-1435-44c5-ab34-a6388d794551" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.082336] env[63293]: DEBUG oslo_concurrency.lockutils [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "f9ccb216-1435-44c5-ab34-a6388d794551" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.082336] env[63293]: DEBUG oslo_concurrency.lockutils [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "f9ccb216-1435-44c5-ab34-a6388d794551-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.082336] env[63293]: DEBUG oslo_concurrency.lockutils [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "f9ccb216-1435-44c5-ab34-a6388d794551-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.082479] env[63293]: DEBUG oslo_concurrency.lockutils [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "f9ccb216-1435-44c5-ab34-a6388d794551-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.087167] env[63293]: INFO nova.compute.manager [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Terminating instance [ 1009.089715] env[63293]: DEBUG nova.compute.manager [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1009.089715] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.090709] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e1614d-259e-4963-a03a-c8b3ff01020f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.098077] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.098362] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8672d69-1d90-4a7d-8a87-71a25b98379b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.105194] env[63293]: DEBUG oslo_vmware.api [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 1009.105194] env[63293]: value = "task-1328100" [ 1009.105194] env[63293]: _type = "Task" [ 1009.105194] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.114289] env[63293]: DEBUG oslo_vmware.api [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328100, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.304209] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.304209] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.304209] env[63293]: DEBUG nova.network.neutron [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1009.363318] env[63293]: DEBUG oslo_concurrency.lockutils [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "71154ca4-ea2c-4187-8319-0bfb9b289eab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.363620] env[63293]: DEBUG oslo_concurrency.lockutils [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "71154ca4-ea2c-4187-8319-0bfb9b289eab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.363841] env[63293]: DEBUG oslo_concurrency.lockutils [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "71154ca4-ea2c-4187-8319-0bfb9b289eab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.364078] env[63293]: DEBUG oslo_concurrency.lockutils [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "71154ca4-ea2c-4187-8319-0bfb9b289eab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.364265] env[63293]: DEBUG oslo_concurrency.lockutils [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "71154ca4-ea2c-4187-8319-0bfb9b289eab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.367801] env[63293]: INFO nova.compute.manager [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Terminating instance [ 1009.369855] env[63293]: DEBUG nova.compute.manager [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1009.370059] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.370952] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b93d0a7-131b-49e6-b49f-280830f1c11c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.379353] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.379644] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1966ae81-48b3-4ce6-a0ac-d9d7a9572693 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.388043] env[63293]: DEBUG oslo_vmware.api [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 1009.388043] env[63293]: value = "task-1328101" [ 1009.388043] env[63293]: _type = "Task" [ 1009.388043] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.391988] env[63293]: DEBUG nova.scheduler.client.report [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.402632] env[63293]: DEBUG oslo_vmware.api [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328101, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.499838] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.553875] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328099, 'name': ReconfigVM_Task, 'duration_secs': 0.40889} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.553875] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance 'd605e066-806a-4aac-9af6-f79d542ab2e3' progress to 33 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1009.599640] env[63293]: INFO nova.network.neutron [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Port c2fa0388-8ddf-497f-8200-903c0fde5a7b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1009.600069] env[63293]: DEBUG nova.network.neutron [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updating instance_info_cache with network_info: [{"id": "5d29fa7c-6329-4700-b988-ac08592d4df7", "address": "fa:16:3e:56:b9:c4", "network": {"id": "908b99b3-6031-4de0-86a4-09a9cbe355a5", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-103315323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "866b347100454019b07f63922b995bb6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d29fa7c-63", "ovs_interfaceid": "5d29fa7c-6329-4700-b988-ac08592d4df7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.616223] env[63293]: DEBUG oslo_vmware.api [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328100, 'name': PowerOffVM_Task, 'duration_secs': 0.368556} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.616508] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1009.616693] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1009.616952] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-302b60b8-2d21-4efd-9c76-9fee044ef8b1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.695264] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1009.695533] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1009.695753] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Deleting the datastore file [datastore2] f9ccb216-1435-44c5-ab34-a6388d794551 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.696193] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c10ffc1-5b0c-4362-b93b-83966e4f0dd4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.708687] env[63293]: DEBUG oslo_vmware.api [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 1009.708687] env[63293]: value = "task-1328103" [ 1009.708687] env[63293]: _type = "Task" [ 1009.708687] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.721462] env[63293]: DEBUG oslo_vmware.api [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328103, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.838834] env[63293]: DEBUG nova.network.neutron [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1009.897763] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.256s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.904091] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.862s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.906397] env[63293]: INFO nova.compute.claims [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1009.908964] env[63293]: DEBUG oslo_vmware.api [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328101, 'name': PowerOffVM_Task, 'duration_secs': 0.261785} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.909531] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1009.909716] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1009.910012] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-108738bb-290e-4e42-b348-6432533757bd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.980995] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1009.981275] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1009.981485] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleting the datastore file [datastore1] 71154ca4-ea2c-4187-8319-0bfb9b289eab {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.981741] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-911a4ce5-ea10-4ba2-881b-e40dcf3e567d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.989537] env[63293]: DEBUG oslo_vmware.api [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for the task: (returnval){ [ 1009.989537] env[63293]: value = "task-1328105" [ 1009.989537] env[63293]: _type = "Task" [ 1009.989537] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.001597] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328098, 'name': ReconfigVM_Task, 'duration_secs': 0.715987} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.004678] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50/2bf4a8a9-c46d-4512-96b7-b34b1bf69f50.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1010.005350] env[63293]: DEBUG oslo_vmware.api [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328105, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.005564] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6987d14-64d7-400d-9ec8-40abc0896502 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.013702] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for the task: (returnval){ [ 1010.013702] env[63293]: value = "task-1328106" [ 1010.013702] env[63293]: _type = "Task" [ 1010.013702] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.023359] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328106, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.038373] env[63293]: DEBUG nova.network.neutron [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance_info_cache with network_info: [{"id": "c589de77-25fc-43f7-9480-e6c97bbda93d", "address": "fa:16:3e:4a:ef:d2", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc589de77-25", "ovs_interfaceid": "c589de77-25fc-43f7-9480-e6c97bbda93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.044469] env[63293]: DEBUG nova.network.neutron [-] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.061985] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1010.062303] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1010.062499] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1010.062740] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1010.062977] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1010.063079] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1010.063283] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1010.063459] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1010.063625] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1010.063888] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1010.063954] env[63293]: DEBUG nova.virt.hardware [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1010.069638] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1010.071842] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2e0751b-af4f-45d1-b23c-cc0e50771bac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.092571] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 1010.092571] env[63293]: value = "task-1328107" [ 1010.092571] env[63293]: _type = "Task" [ 1010.092571] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.101660] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328107, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.103269] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Releasing lock "refresh_cache-f9ccb216-1435-44c5-ab34-a6388d794551" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.219075] env[63293]: DEBUG oslo_vmware.api [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328103, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170779} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.219351] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1010.219542] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1010.219745] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1010.219900] env[63293]: INFO nova.compute.manager [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1010.220158] env[63293]: DEBUG oslo.service.loopingcall [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.220384] env[63293]: DEBUG nova.compute.manager [-] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1010.220484] env[63293]: DEBUG nova.network.neutron [-] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1010.423108] env[63293]: DEBUG oslo_concurrency.lockutils [None req-30162edb-2248-4fea-a949-df7a45873481 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "f429869f-9e9f-443e-a980-67a80c3bd799" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 27.761s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.423108] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "f429869f-9e9f-443e-a980-67a80c3bd799" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 7.084s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.423108] env[63293]: INFO nova.compute.manager [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Unshelving [ 1010.504533] env[63293]: DEBUG oslo_vmware.api [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Task: {'id': task-1328105, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.202802} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.504858] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1010.505078] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1010.505564] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1010.505753] env[63293]: INFO nova.compute.manager [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1010.506014] env[63293]: DEBUG oslo.service.loopingcall [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.506220] env[63293]: DEBUG nova.compute.manager [-] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1010.506495] env[63293]: DEBUG nova.network.neutron [-] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1010.523052] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328106, 'name': Rename_Task, 'duration_secs': 0.156365} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.523330] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1010.523583] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18076fa3-21bb-4ee3-b7da-d3463fb92287 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.531679] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for the task: (returnval){ [ 1010.531679] env[63293]: value = "task-1328108" [ 1010.531679] env[63293]: _type = "Task" [ 1010.531679] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.539814] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328108, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.541340] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.541623] env[63293]: DEBUG nova.compute.manager [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Instance network_info: |[{"id": "c589de77-25fc-43f7-9480-e6c97bbda93d", "address": "fa:16:3e:4a:ef:d2", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc589de77-25", "ovs_interfaceid": "c589de77-25fc-43f7-9480-e6c97bbda93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1010.541999] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:ef:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c589de77-25fc-43f7-9480-e6c97bbda93d', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1010.549674] env[63293]: DEBUG oslo.service.loopingcall [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.550076] env[63293]: INFO nova.compute.manager [-] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Took 1.51 seconds to deallocate network for instance. [ 1010.550347] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1010.552293] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d4d9b47-c712-4496-9ab0-ab2dc8da4fcc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.580301] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1010.580301] env[63293]: value = "task-1328109" [ 1010.580301] env[63293]: _type = "Task" [ 1010.580301] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.592330] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328109, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.602182] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328107, 'name': ReconfigVM_Task, 'duration_secs': 0.379444} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.602473] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1010.603288] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d50fd8-1e26-42a8-a4ae-ff4ce0b3726b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.606863] env[63293]: DEBUG oslo_concurrency.lockutils [None req-fc66ad87-0515-4aab-9dd0-9417ee1a4b47 tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "interface-f9ccb216-1435-44c5-ab34-a6388d794551-c2fa0388-8ddf-497f-8200-903c0fde5a7b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.076s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.625923] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] d605e066-806a-4aac-9af6-f79d542ab2e3/d605e066-806a-4aac-9af6-f79d542ab2e3.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1010.627400] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cecc8f7a-291d-41d6-a87d-31b4d0406437 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.646669] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 1010.646669] env[63293]: value = "task-1328110" [ 1010.646669] env[63293]: _type = "Task" [ 1010.646669] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.657970] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328110, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.823906] env[63293]: DEBUG nova.compute.manager [req-96ccc4a2-12da-4015-a539-0e52091a89c4 req-9b2a4a7f-fa55-492d-92ea-c1c58129816a service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Received event network-changed-c589de77-25fc-43f7-9480-e6c97bbda93d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.828034] env[63293]: DEBUG nova.compute.manager [req-96ccc4a2-12da-4015-a539-0e52091a89c4 req-9b2a4a7f-fa55-492d-92ea-c1c58129816a service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Refreshing instance network info cache due to event network-changed-c589de77-25fc-43f7-9480-e6c97bbda93d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1010.828034] env[63293]: DEBUG oslo_concurrency.lockutils [req-96ccc4a2-12da-4015-a539-0e52091a89c4 req-9b2a4a7f-fa55-492d-92ea-c1c58129816a service nova] Acquiring lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.828034] env[63293]: DEBUG oslo_concurrency.lockutils [req-96ccc4a2-12da-4015-a539-0e52091a89c4 req-9b2a4a7f-fa55-492d-92ea-c1c58129816a service nova] Acquired lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.828034] env[63293]: DEBUG nova.network.neutron [req-96ccc4a2-12da-4015-a539-0e52091a89c4 req-9b2a4a7f-fa55-492d-92ea-c1c58129816a service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Refreshing network info cache for port c589de77-25fc-43f7-9480-e6c97bbda93d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1011.042669] env[63293]: DEBUG oslo_vmware.api [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328108, 'name': PowerOnVM_Task, 'duration_secs': 0.496271} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.042992] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1011.044311] env[63293]: INFO nova.compute.manager [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Took 7.69 seconds to spawn the instance on the hypervisor. [ 1011.044311] env[63293]: DEBUG nova.compute.manager [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1011.044991] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f886c057-7525-4847-a044-c1324d721731 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.078964] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.091289] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328109, 'name': CreateVM_Task, 'duration_secs': 0.337681} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.093901] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1011.095955] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.096187] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.096544] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1011.097082] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-248c167b-20d6-4f2e-b209-1093d86f9ea7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.105223] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1011.105223] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e55d33-1dfa-4183-d232-ca48920af965" [ 1011.105223] env[63293]: _type = "Task" [ 1011.105223] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.111209] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af82c3f0-1d9e-499e-9e1f-12f8bfbf229b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.119863] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e55d33-1dfa-4183-d232-ca48920af965, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.124013] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf757174-3c7a-43f0-a12a-88fc8de10d07 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.164154] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7dbfce-28bc-4bbf-8039-e7123a178e62 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.173587] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328110, 'name': ReconfigVM_Task, 'duration_secs': 0.301983} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.175871] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Reconfigured VM instance instance-00000060 to attach disk [datastore1] d605e066-806a-4aac-9af6-f79d542ab2e3/d605e066-806a-4aac-9af6-f79d542ab2e3.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1011.176219] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance 'd605e066-806a-4aac-9af6-f79d542ab2e3' progress to 50 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1011.184925] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c9bcea-34cd-488d-abee-15f4432e67b7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.204023] env[63293]: DEBUG nova.compute.provider_tree [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.448515] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.567013] env[63293]: INFO nova.compute.manager [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Took 12.59 seconds to build instance. [ 1011.616021] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e55d33-1dfa-4183-d232-ca48920af965, 'name': SearchDatastore_Task, 'duration_secs': 0.015672} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.616127] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.616281] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1011.616514] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.616830] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.616830] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1011.617107] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-413f754c-2b9f-4308-964f-3f23119f0ad0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.621601] env[63293]: DEBUG nova.network.neutron [req-96ccc4a2-12da-4015-a539-0e52091a89c4 req-9b2a4a7f-fa55-492d-92ea-c1c58129816a service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updated VIF entry in instance network info cache for port c589de77-25fc-43f7-9480-e6c97bbda93d. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1011.621953] env[63293]: DEBUG nova.network.neutron [req-96ccc4a2-12da-4015-a539-0e52091a89c4 req-9b2a4a7f-fa55-492d-92ea-c1c58129816a service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance_info_cache with network_info: [{"id": "c589de77-25fc-43f7-9480-e6c97bbda93d", "address": "fa:16:3e:4a:ef:d2", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc589de77-25", "ovs_interfaceid": "c589de77-25fc-43f7-9480-e6c97bbda93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.631185] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1011.631370] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1011.632614] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da17ead7-a58a-4abf-aafa-075a0cdb3889 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.638789] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1011.638789] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528d8572-1109-bfd5-2bf7-07e1c95af649" [ 1011.638789] env[63293]: _type = "Task" [ 1011.638789] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.647224] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528d8572-1109-bfd5-2bf7-07e1c95af649, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.694839] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361a9f03-3a10-4c3a-b9d5-b12d6ca541b2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.718531] env[63293]: DEBUG nova.scheduler.client.report [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.722712] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6bb762-50c0-4179-a698-172f412de8c5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.743008] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance 'd605e066-806a-4aac-9af6-f79d542ab2e3' progress to 67 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1011.799313] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "97da107a-2e15-4da2-9197-7f85d960ff27" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.799594] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.806026] env[63293]: DEBUG nova.network.neutron [-] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.048739] env[63293]: DEBUG nova.network.neutron [-] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.068945] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ec282957-717d-465e-a997-f1fba1601ac5 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.096s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.124878] env[63293]: DEBUG oslo_concurrency.lockutils [req-96ccc4a2-12da-4015-a539-0e52091a89c4 req-9b2a4a7f-fa55-492d-92ea-c1c58129816a service nova] Releasing lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.124878] env[63293]: DEBUG nova.compute.manager [req-96ccc4a2-12da-4015-a539-0e52091a89c4 req-9b2a4a7f-fa55-492d-92ea-c1c58129816a service nova] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Received event network-vif-deleted-12ae71a2-a302-4223-8d65-630cde456181 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.150628] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528d8572-1109-bfd5-2bf7-07e1c95af649, 'name': SearchDatastore_Task, 'duration_secs': 0.034864} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.151834] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1a60f91-b428-4f68-be0a-b0b37007655f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.156999] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1012.156999] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52315393-ddf8-541c-6e19-527ea9734e32" [ 1012.156999] env[63293]: _type = "Task" [ 1012.156999] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.166567] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52315393-ddf8-541c-6e19-527ea9734e32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.226996] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.227603] env[63293]: DEBUG nova.compute.manager [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1012.231542] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.155s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.231542] env[63293]: DEBUG nova.objects.instance [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lazy-loading 'resources' on Instance uuid bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.298718] env[63293]: DEBUG nova.network.neutron [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Port 83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970 binding to destination host cpu-1 is already ACTIVE {{(pid=63293) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1012.305597] env[63293]: DEBUG nova.compute.manager [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1012.308304] env[63293]: INFO nova.compute.manager [-] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Took 1.80 seconds to deallocate network for instance. [ 1012.553674] env[63293]: INFO nova.compute.manager [-] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Took 2.33 seconds to deallocate network for instance. [ 1012.668809] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52315393-ddf8-541c-6e19-527ea9734e32, 'name': SearchDatastore_Task, 'duration_secs': 0.036846} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.669117] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.669354] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] bea1ea16-c224-4a0e-86ed-c2aaf920d3e1/bea1ea16-c224-4a0e-86ed-c2aaf920d3e1.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1012.669623] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b6d22f7d-9a6e-497f-b39d-08fb472d1e5b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.677737] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1012.677737] env[63293]: value = "task-1328111" [ 1012.677737] env[63293]: _type = "Task" [ 1012.677737] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.686546] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328111, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.733513] env[63293]: DEBUG nova.compute.utils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1012.737435] env[63293]: DEBUG nova.compute.manager [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1012.737628] env[63293]: DEBUG nova.network.neutron [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1012.792727] env[63293]: DEBUG nova.policy [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ecbac9f9fca4b42a14c3ca0d26e5d34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'affba61d6a2846b38666544bc2c25db5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 1012.816931] env[63293]: DEBUG oslo_concurrency.lockutils [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.827789] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.870786] env[63293]: DEBUG nova.compute.manager [req-8bb53b48-e028-4894-a229-e5d228b23c7b req-91fe46b5-9b4d-4570-9c8c-ee6ca7fdcbfe service nova] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Received event network-vif-deleted-69771278-cd61-43d4-b53c-76544685f282 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.870786] env[63293]: DEBUG nova.compute.manager [req-8bb53b48-e028-4894-a229-e5d228b23c7b req-91fe46b5-9b4d-4570-9c8c-ee6ca7fdcbfe service nova] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Received event network-vif-deleted-5d29fa7c-6329-4700-b988-ac08592d4df7 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.870786] env[63293]: DEBUG nova.compute.manager [req-8bb53b48-e028-4894-a229-e5d228b23c7b req-91fe46b5-9b4d-4570-9c8c-ee6ca7fdcbfe service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Received event network-changed-8147f046-74b1-4a9e-b437-02f3a5ee8fe3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.870786] env[63293]: DEBUG nova.compute.manager [req-8bb53b48-e028-4894-a229-e5d228b23c7b req-91fe46b5-9b4d-4570-9c8c-ee6ca7fdcbfe service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Refreshing instance network info cache due to event network-changed-8147f046-74b1-4a9e-b437-02f3a5ee8fe3. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1012.871979] env[63293]: DEBUG oslo_concurrency.lockutils [req-8bb53b48-e028-4894-a229-e5d228b23c7b req-91fe46b5-9b4d-4570-9c8c-ee6ca7fdcbfe service nova] Acquiring lock "refresh_cache-2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.871979] env[63293]: DEBUG oslo_concurrency.lockutils [req-8bb53b48-e028-4894-a229-e5d228b23c7b req-91fe46b5-9b4d-4570-9c8c-ee6ca7fdcbfe service nova] Acquired lock "refresh_cache-2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.871979] env[63293]: DEBUG nova.network.neutron [req-8bb53b48-e028-4894-a229-e5d228b23c7b req-91fe46b5-9b4d-4570-9c8c-ee6ca7fdcbfe service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Refreshing network info cache for port 8147f046-74b1-4a9e-b437-02f3a5ee8fe3 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1012.920068] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85d25db-79ca-444a-afa6-6d2b1f010bdd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.928673] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af78a21-b67f-416f-9f0d-dc9c756a143c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.961876] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c0ca7a-511c-426a-a7e4-8f3e6e7e5579 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.970202] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a8eee1-9010-4e1b-8b73-fcff6295e815 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.986560] env[63293]: DEBUG nova.compute.provider_tree [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.060229] env[63293]: DEBUG oslo_concurrency.lockutils [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.097277] env[63293]: DEBUG nova.network.neutron [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Successfully created port: 3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1013.189201] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328111, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.240980] env[63293]: DEBUG nova.compute.manager [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1013.321684] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "d605e066-806a-4aac-9af6-f79d542ab2e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.321823] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.321949] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.490496] env[63293]: DEBUG nova.scheduler.client.report [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.652059] env[63293]: DEBUG nova.network.neutron [req-8bb53b48-e028-4894-a229-e5d228b23c7b req-91fe46b5-9b4d-4570-9c8c-ee6ca7fdcbfe service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Updated VIF entry in instance network info cache for port 8147f046-74b1-4a9e-b437-02f3a5ee8fe3. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1013.652377] env[63293]: DEBUG nova.network.neutron [req-8bb53b48-e028-4894-a229-e5d228b23c7b req-91fe46b5-9b4d-4570-9c8c-ee6ca7fdcbfe service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Updating instance_info_cache with network_info: [{"id": "8147f046-74b1-4a9e-b437-02f3a5ee8fe3", "address": "fa:16:3e:c2:a4:a0", "network": {"id": "39a6f586-5383-45ed-999b-db33c8f4d9f3", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-884746607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97cb1a48336448ad866206ab157778c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8147f046-74", "ovs_interfaceid": "8147f046-74b1-4a9e-b437-02f3a5ee8fe3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.689661] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328111, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.560927} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.690063] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] bea1ea16-c224-4a0e-86ed-c2aaf920d3e1/bea1ea16-c224-4a0e-86ed-c2aaf920d3e1.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1013.690418] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.690758] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e4acd0a5-a8a5-40e9-8005-ac5fa871a6e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.697559] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1013.697559] env[63293]: value = "task-1328112" [ 1013.697559] env[63293]: _type = "Task" [ 1013.697559] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.706221] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328112, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.004492] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.774s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.006841] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.558s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.007084] env[63293]: DEBUG nova.objects.instance [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lazy-loading 'pci_requests' on Instance uuid f429869f-9e9f-443e-a980-67a80c3bd799 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.030734] env[63293]: INFO nova.scheduler.client.report [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted allocations for instance bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7 [ 1014.154772] env[63293]: DEBUG oslo_concurrency.lockutils [req-8bb53b48-e028-4894-a229-e5d228b23c7b req-91fe46b5-9b4d-4570-9c8c-ee6ca7fdcbfe service nova] Releasing lock "refresh_cache-2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.207250] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328112, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069836} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.207531] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1014.208345] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67f5fc3-05f4-4a84-9327-909f19b0e465 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.232018] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] bea1ea16-c224-4a0e-86ed-c2aaf920d3e1/bea1ea16-c224-4a0e-86ed-c2aaf920d3e1.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.232324] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e076297a-01b1-4fc4-8273-07522e7665e0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.251975] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1014.251975] env[63293]: value = "task-1328113" [ 1014.251975] env[63293]: _type = "Task" [ 1014.251975] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.256062] env[63293]: DEBUG nova.compute.manager [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1014.264097] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328113, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.280828] env[63293]: DEBUG nova.virt.hardware [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.281172] env[63293]: DEBUG nova.virt.hardware [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.281351] env[63293]: DEBUG nova.virt.hardware [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.281690] env[63293]: DEBUG nova.virt.hardware [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.281896] env[63293]: DEBUG nova.virt.hardware [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.282099] env[63293]: DEBUG nova.virt.hardware [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1014.282358] env[63293]: DEBUG nova.virt.hardware [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1014.282564] env[63293]: DEBUG nova.virt.hardware [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1014.282787] env[63293]: DEBUG nova.virt.hardware [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1014.283046] env[63293]: DEBUG nova.virt.hardware [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1014.283279] env[63293]: DEBUG nova.virt.hardware [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.284135] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90bc90d6-f4b4-4301-9723-d0293e48ba1f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.292566] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d249f1-984f-4c11-88a5-e6ac1bad39d7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.371275] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.371503] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.371694] env[63293]: DEBUG nova.network.neutron [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1014.511642] env[63293]: DEBUG nova.objects.instance [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lazy-loading 'numa_topology' on Instance uuid f429869f-9e9f-443e-a980-67a80c3bd799 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.538151] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5d0bf151-f277-43b8-b131-3c19e4b01bee tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.656s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.762685] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328113, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.814187] env[63293]: DEBUG nova.network.neutron [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Successfully updated port: 3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1014.896125] env[63293]: DEBUG nova.compute.manager [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Received event network-vif-plugged-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.896408] env[63293]: DEBUG oslo_concurrency.lockutils [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] Acquiring lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.896660] env[63293]: DEBUG oslo_concurrency.lockutils [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.896875] env[63293]: DEBUG oslo_concurrency.lockutils [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.897103] env[63293]: DEBUG nova.compute.manager [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] No waiting events found dispatching network-vif-plugged-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.897400] env[63293]: WARNING nova.compute.manager [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Received unexpected event network-vif-plugged-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 for instance with vm_state building and task_state spawning. [ 1014.897513] env[63293]: DEBUG nova.compute.manager [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Received event network-changed-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.897723] env[63293]: DEBUG nova.compute.manager [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Refreshing instance network info cache due to event network-changed-3313c3a6-17f4-4c97-b19f-9edfd6eb4099. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1014.897933] env[63293]: DEBUG oslo_concurrency.lockutils [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] Acquiring lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.898130] env[63293]: DEBUG oslo_concurrency.lockutils [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] Acquired lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.898349] env[63293]: DEBUG nova.network.neutron [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Refreshing network info cache for port 3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.013718] env[63293]: INFO nova.compute.claims [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1015.144176] env[63293]: DEBUG nova.network.neutron [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance_info_cache with network_info: [{"id": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "address": "fa:16:3e:c2:85:7d", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83a30fb0-a4", "ovs_interfaceid": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.263704] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328113, 'name': ReconfigVM_Task, 'duration_secs': 0.904558} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.264043] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfigured VM instance instance-00000064 to attach disk [datastore2] bea1ea16-c224-4a0e-86ed-c2aaf920d3e1/bea1ea16-c224-4a0e-86ed-c2aaf920d3e1.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1015.264702] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c823415c-acd3-4f7e-b432-b3b6e05c0a02 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.271528] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1015.271528] env[63293]: value = "task-1328114" [ 1015.271528] env[63293]: _type = "Task" [ 1015.271528] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.278962] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328114, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.316432] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.359494] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.359788] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.430500] env[63293]: DEBUG nova.network.neutron [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1015.515360] env[63293]: DEBUG nova.network.neutron [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.646979] env[63293]: DEBUG oslo_concurrency.lockutils [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.781701] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328114, 'name': Rename_Task, 'duration_secs': 0.141878} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.782045] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1015.782249] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e56db57-fe97-42fc-a2b8-743ba677f213 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.788058] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1015.788058] env[63293]: value = "task-1328115" [ 1015.788058] env[63293]: _type = "Task" [ 1015.788058] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.795123] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328115, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.862398] env[63293]: DEBUG nova.compute.manager [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1016.021202] env[63293]: DEBUG oslo_concurrency.lockutils [req-71373ba0-684d-46d1-9815-27728c3aa888 req-7d30450c-faa6-4b4c-80d4-548b3c398834 service nova] Releasing lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.025422] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.025637] env[63293]: DEBUG nova.network.neutron [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1016.171127] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f5d270-b1ae-46bd-a5fd-08394fd75f3d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.177244] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f030e51f-919e-44d5-aede-1481ff33a6db {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.210298] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698ad4a1-18cd-4904-86bc-2eb87b822c51 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.216399] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e6f786-b751-4136-9919-b2d1985d5f42 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.223923] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance 'd605e066-806a-4aac-9af6-f79d542ab2e3' progress to 83 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1016.253203] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e2b237-db72-4086-8ce6-33961ae5e035 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.260993] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d4dc5c-8902-444b-93b4-d08ffd251dea {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.274920] env[63293]: DEBUG nova.compute.provider_tree [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.296977] env[63293]: DEBUG oslo_vmware.api [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328115, 'name': PowerOnVM_Task, 'duration_secs': 0.502459} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.297215] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1016.297417] env[63293]: INFO nova.compute.manager [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Took 8.39 seconds to spawn the instance on the hypervisor. [ 1016.297599] env[63293]: DEBUG nova.compute.manager [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.298393] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f5c442-560e-4b22-b273-e985305e156a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.380476] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.559517] env[63293]: DEBUG nova.network.neutron [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1016.703563] env[63293]: DEBUG nova.network.neutron [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updating instance_info_cache with network_info: [{"id": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "address": "fa:16:3e:85:83:c3", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3313c3a6-17", "ovs_interfaceid": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.754804] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1016.755178] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42f013a8-f780-4303-8924-a1fbfef6b5ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.762271] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 1016.762271] env[63293]: value = "task-1328116" [ 1016.762271] env[63293]: _type = "Task" [ 1016.762271] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.769928] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328116, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.777874] env[63293]: DEBUG nova.scheduler.client.report [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.816459] env[63293]: INFO nova.compute.manager [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Took 16.24 seconds to build instance. [ 1017.207039] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.207413] env[63293]: DEBUG nova.compute.manager [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Instance network_info: |[{"id": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "address": "fa:16:3e:85:83:c3", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3313c3a6-17", "ovs_interfaceid": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1017.207871] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:83:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24144f5a-050a-4f1e-8d8c-774dc16dc791', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3313c3a6-17f4-4c97-b19f-9edfd6eb4099', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.215486] env[63293]: DEBUG oslo.service.loopingcall [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.215731] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1017.215970] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9cfa55e1-11ec-4745-8a47-a68190f6b707 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.235714] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.235714] env[63293]: value = "task-1328117" [ 1017.235714] env[63293]: _type = "Task" [ 1017.235714] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.244951] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328117, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.272110] env[63293]: DEBUG oslo_vmware.api [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328116, 'name': PowerOnVM_Task, 'duration_secs': 0.388759} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.272394] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1017.272583] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-8f27203f-4680-486e-bd15-c72dd702643b tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance 'd605e066-806a-4aac-9af6-f79d542ab2e3' progress to 100 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1017.282541] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.276s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.284417] env[63293]: DEBUG oslo_concurrency.lockutils [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.468s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.284652] env[63293]: DEBUG nova.objects.instance [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lazy-loading 'resources' on Instance uuid 71154ca4-ea2c-4187-8319-0bfb9b289eab {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.319221] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b709d958-694c-4179-add5-f7e76d87db9a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.755s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.327379] env[63293]: INFO nova.network.neutron [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updating port 4c817793-6a1c-4d84-ac42-0487913a257b with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1017.747421] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328117, 'name': CreateVM_Task, 'duration_secs': 0.286292} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.747645] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1017.748550] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.748628] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.749020] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1017.749326] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b2d1e91-ce8f-4d66-89c5-b2238a21d207 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.754425] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1017.754425] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523c78b5-d77f-d458-aae0-251375d0cc09" [ 1017.754425] env[63293]: _type = "Task" [ 1017.754425] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.763328] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523c78b5-d77f-d458-aae0-251375d0cc09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.902317] env[63293]: DEBUG nova.compute.manager [req-e950e1a5-8f0a-4d5f-b371-fe3026ced9cf req-17785cb6-1bdf-4dee-9e74-7ebd8b14cecb service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Received event network-changed-c589de77-25fc-43f7-9480-e6c97bbda93d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.902570] env[63293]: DEBUG nova.compute.manager [req-e950e1a5-8f0a-4d5f-b371-fe3026ced9cf req-17785cb6-1bdf-4dee-9e74-7ebd8b14cecb service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Refreshing instance network info cache due to event network-changed-c589de77-25fc-43f7-9480-e6c97bbda93d. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1017.902720] env[63293]: DEBUG oslo_concurrency.lockutils [req-e950e1a5-8f0a-4d5f-b371-fe3026ced9cf req-17785cb6-1bdf-4dee-9e74-7ebd8b14cecb service nova] Acquiring lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.902863] env[63293]: DEBUG oslo_concurrency.lockutils [req-e950e1a5-8f0a-4d5f-b371-fe3026ced9cf req-17785cb6-1bdf-4dee-9e74-7ebd8b14cecb service nova] Acquired lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.903031] env[63293]: DEBUG nova.network.neutron [req-e950e1a5-8f0a-4d5f-b371-fe3026ced9cf req-17785cb6-1bdf-4dee-9e74-7ebd8b14cecb service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Refreshing network info cache for port c589de77-25fc-43f7-9480-e6c97bbda93d {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1017.943490] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb52a620-2481-4e16-bb2b-49808badc96d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.951139] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6255a52c-b089-4338-94f9-2b4148c6c0ae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.982966] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a313d9-333c-43a2-bd3e-44c936625e30 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.991025] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff1af16-05f9-43d3-afcb-2c42c3b157e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.004708] env[63293]: DEBUG nova.compute.provider_tree [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.264123] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]523c78b5-d77f-d458-aae0-251375d0cc09, 'name': SearchDatastore_Task, 'duration_secs': 0.009478} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.264444] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.264681] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.264918] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.265086] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.265267] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.265526] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc1e9d04-555a-49be-b7bf-50c78e599322 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.274276] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.274456] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.275154] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18bb4a46-d989-4851-bcce-646e1e454ed8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.283747] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1018.283747] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5221db38-56d7-01ac-ed7a-8f11860a3857" [ 1018.283747] env[63293]: _type = "Task" [ 1018.283747] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.292373] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5221db38-56d7-01ac-ed7a-8f11860a3857, 'name': SearchDatastore_Task, 'duration_secs': 0.008777} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.293085] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79ba2885-8cbe-4013-af71-b4b7a3db925c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.297468] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1018.297468] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5210618f-484c-5b28-8ffd-ae98f0e844aa" [ 1018.297468] env[63293]: _type = "Task" [ 1018.297468] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.304827] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5210618f-484c-5b28-8ffd-ae98f0e844aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.508295] env[63293]: DEBUG nova.scheduler.client.report [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.719666] env[63293]: DEBUG nova.network.neutron [req-e950e1a5-8f0a-4d5f-b371-fe3026ced9cf req-17785cb6-1bdf-4dee-9e74-7ebd8b14cecb service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updated VIF entry in instance network info cache for port c589de77-25fc-43f7-9480-e6c97bbda93d. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1018.720108] env[63293]: DEBUG nova.network.neutron [req-e950e1a5-8f0a-4d5f-b371-fe3026ced9cf req-17785cb6-1bdf-4dee-9e74-7ebd8b14cecb service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance_info_cache with network_info: [{"id": "c589de77-25fc-43f7-9480-e6c97bbda93d", "address": "fa:16:3e:4a:ef:d2", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc589de77-25", "ovs_interfaceid": "c589de77-25fc-43f7-9480-e6c97bbda93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.814617] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5210618f-484c-5b28-8ffd-ae98f0e844aa, 'name': SearchDatastore_Task, 'duration_secs': 0.008549} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.815086] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.815495] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d/97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1018.815866] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4cadda0b-099d-4a64-8807-c17a2f4769f0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.823847] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1018.823847] env[63293]: value = "task-1328118" [ 1018.823847] env[63293]: _type = "Task" [ 1018.823847] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.833791] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328118, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.860295] env[63293]: DEBUG nova.compute.manager [req-f10202d7-e9ef-436e-a3cc-d3c460abe2cb req-6b223aa8-ecf6-4207-92d1-55f9f75c9282 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Received event network-vif-plugged-4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.860575] env[63293]: DEBUG oslo_concurrency.lockutils [req-f10202d7-e9ef-436e-a3cc-d3c460abe2cb req-6b223aa8-ecf6-4207-92d1-55f9f75c9282 service nova] Acquiring lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.863109] env[63293]: DEBUG oslo_concurrency.lockutils [req-f10202d7-e9ef-436e-a3cc-d3c460abe2cb req-6b223aa8-ecf6-4207-92d1-55f9f75c9282 service nova] Lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.863109] env[63293]: DEBUG oslo_concurrency.lockutils [req-f10202d7-e9ef-436e-a3cc-d3c460abe2cb req-6b223aa8-ecf6-4207-92d1-55f9f75c9282 service nova] Lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.863109] env[63293]: DEBUG nova.compute.manager [req-f10202d7-e9ef-436e-a3cc-d3c460abe2cb req-6b223aa8-ecf6-4207-92d1-55f9f75c9282 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] No waiting events found dispatching network-vif-plugged-4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1018.863109] env[63293]: WARNING nova.compute.manager [req-f10202d7-e9ef-436e-a3cc-d3c460abe2cb req-6b223aa8-ecf6-4207-92d1-55f9f75c9282 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Received unexpected event network-vif-plugged-4c817793-6a1c-4d84-ac42-0487913a257b for instance with vm_state shelved_offloaded and task_state spawning. [ 1018.935778] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.935778] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.936280] env[63293]: DEBUG nova.network.neutron [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1019.013387] env[63293]: DEBUG oslo_concurrency.lockutils [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.729s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.016737] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.189s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.018583] env[63293]: INFO nova.compute.claims [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1019.044673] env[63293]: INFO nova.scheduler.client.report [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Deleted allocations for instance 71154ca4-ea2c-4187-8319-0bfb9b289eab [ 1019.224317] env[63293]: DEBUG oslo_concurrency.lockutils [req-e950e1a5-8f0a-4d5f-b371-fe3026ced9cf req-17785cb6-1bdf-4dee-9e74-7ebd8b14cecb service nova] Releasing lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.339486] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328118, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485236} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.339641] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d/97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1019.339993] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1019.340407] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0acfbc5-b720-4772-84a3-5aaa03b1e906 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.348752] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1019.348752] env[63293]: value = "task-1328119" [ 1019.348752] env[63293]: _type = "Task" [ 1019.348752] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.358065] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328119, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.445057] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "d605e066-806a-4aac-9af6-f79d542ab2e3" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.445368] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.445519] env[63293]: DEBUG nova.compute.manager [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Going to confirm migration 4 {{(pid=63293) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1019.553861] env[63293]: DEBUG oslo_concurrency.lockutils [None req-778e010a-8ca7-42d7-9946-458225e196be tempest-ServerDiskConfigTestJSON-139540904 tempest-ServerDiskConfigTestJSON-139540904-project-member] Lock "71154ca4-ea2c-4187-8319-0bfb9b289eab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.190s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.646597] env[63293]: DEBUG nova.network.neutron [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updating instance_info_cache with network_info: [{"id": "4c817793-6a1c-4d84-ac42-0487913a257b", "address": "fa:16:3e:18:84:4b", "network": {"id": "8af42ccf-7ab8-424b-a290-8383c292ef85", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1850566934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eec85c5992d64d62a90e290a1aa2d441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c817793-6a", "ovs_interfaceid": "4c817793-6a1c-4d84-ac42-0487913a257b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.858838] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328119, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060905} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.859132] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1019.859926] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e0f276-96b3-4d37-8f38-5527e2fc2b2f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.882685] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d/97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1019.883038] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9eedfd9-ef9f-470d-941a-3fdc3efe74fa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.902293] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1019.902293] env[63293]: value = "task-1328120" [ 1019.902293] env[63293]: _type = "Task" [ 1019.902293] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.912156] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328120, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.022944] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.023314] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquired lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.023582] env[63293]: DEBUG nova.network.neutron [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1020.023798] env[63293]: DEBUG nova.objects.instance [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lazy-loading 'info_cache' on Instance uuid d605e066-806a-4aac-9af6-f79d542ab2e3 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.149521] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.184534] env[63293]: DEBUG nova.virt.hardware [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='0c70b4f11a2a0274cef2543a09d56c6b',container_format='bare',created_at=2024-10-21T17:41:24Z,direct_url=,disk_format='vmdk',id=dc0bd04e-b8f4-482d-b243-fa28e2f86738,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1898489599-shelved',owner='eec85c5992d64d62a90e290a1aa2d441',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-21T17:41:39Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.184785] env[63293]: DEBUG nova.virt.hardware [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.184949] env[63293]: DEBUG nova.virt.hardware [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.185627] env[63293]: DEBUG nova.virt.hardware [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.185727] env[63293]: DEBUG nova.virt.hardware [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.185872] env[63293]: DEBUG nova.virt.hardware [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.186102] env[63293]: DEBUG nova.virt.hardware [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.186791] env[63293]: DEBUG nova.virt.hardware [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.186791] env[63293]: DEBUG nova.virt.hardware [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.186791] env[63293]: DEBUG nova.virt.hardware [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.186967] env[63293]: DEBUG nova.virt.hardware [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.188265] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c250c40-a00b-4ace-a8ef-7aed72fd7312 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.200253] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2afdae5-89f5-45da-b47b-c1650cf9abf0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.206698] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6c1601-b40a-487f-9e6e-ddde5a4e0fd8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.220829] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:84:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd48f0ef6-34e5-44d4-8baf-4470ed96ce73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c817793-6a1c-4d84-ac42-0487913a257b', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.228980] env[63293]: DEBUG oslo.service.loopingcall [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.231119] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1020.231397] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e9ffc6a-8e81-4270-9746-2874fcb19766 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.246739] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1cfda71-08b6-4c9a-b71d-204021ff4d30 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.280457] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd1bb3f5-1a26-410d-ab1c-5a62babee48d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.283722] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.283722] env[63293]: value = "task-1328121" [ 1020.283722] env[63293]: _type = "Task" [ 1020.283722] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.290636] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae47925-ea33-4732-943e-8144824bb8e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.297357] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328121, 'name': CreateVM_Task} progress is 15%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.308733] env[63293]: DEBUG nova.compute.provider_tree [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.412669] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328120, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.793563] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328121, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.811025] env[63293]: DEBUG nova.scheduler.client.report [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.911532] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328120, 'name': ReconfigVM_Task, 'duration_secs': 0.840406} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.911829] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d/97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1020.912542] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13b211fb-9685-4fe5-bc25-1335706e8280 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.918509] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1020.918509] env[63293]: value = "task-1328122" [ 1020.918509] env[63293]: _type = "Task" [ 1020.918509] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.925925] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328122, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.947158] env[63293]: DEBUG nova.compute.manager [req-57ef31f5-ce1c-4958-b96e-482185cba0a2 req-7ab1efbd-f6aa-4f43-830d-4556e046e43d service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Received event network-changed-4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.947422] env[63293]: DEBUG nova.compute.manager [req-57ef31f5-ce1c-4958-b96e-482185cba0a2 req-7ab1efbd-f6aa-4f43-830d-4556e046e43d service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Refreshing instance network info cache due to event network-changed-4c817793-6a1c-4d84-ac42-0487913a257b. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1020.947607] env[63293]: DEBUG oslo_concurrency.lockutils [req-57ef31f5-ce1c-4958-b96e-482185cba0a2 req-7ab1efbd-f6aa-4f43-830d-4556e046e43d service nova] Acquiring lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.947721] env[63293]: DEBUG oslo_concurrency.lockutils [req-57ef31f5-ce1c-4958-b96e-482185cba0a2 req-7ab1efbd-f6aa-4f43-830d-4556e046e43d service nova] Acquired lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.947890] env[63293]: DEBUG nova.network.neutron [req-57ef31f5-ce1c-4958-b96e-482185cba0a2 req-7ab1efbd-f6aa-4f43-830d-4556e046e43d service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Refreshing network info cache for port 4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1021.294008] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328121, 'name': CreateVM_Task, 'duration_secs': 0.551111} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.294251] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1021.294888] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.295076] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired lock "[datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.295459] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1021.295725] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5a87c24-3ea0-45b5-a833-3ac0f6a70b6c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.300446] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1021.300446] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5258e3ae-4bc0-d4d3-0b47-3be1d98a71c1" [ 1021.300446] env[63293]: _type = "Task" [ 1021.300446] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.309679] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5258e3ae-4bc0-d4d3-0b47-3be1d98a71c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.320886] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.321432] env[63293]: DEBUG nova.compute.manager [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1021.325968] env[63293]: DEBUG oslo_concurrency.lockutils [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.265s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.325968] env[63293]: DEBUG nova.objects.instance [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'resources' on Instance uuid f9ccb216-1435-44c5-ab34-a6388d794551 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.352665] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.352946] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.427876] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328122, 'name': Rename_Task, 'duration_secs': 0.474879} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.428184] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1021.428434] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4aca174-ec2c-4a78-a36b-953a5977f3bc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.435193] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1021.435193] env[63293]: value = "task-1328123" [ 1021.435193] env[63293]: _type = "Task" [ 1021.435193] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.444986] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328123, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.585605] env[63293]: DEBUG nova.network.neutron [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance_info_cache with network_info: [{"id": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "address": "fa:16:3e:c2:85:7d", "network": {"id": "f5af0e6d-f5ff-4f23-a7db-cc586dbd186e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-791337968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff8479fc251544e9b4618deccf992754", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83a30fb0-a4", "ovs_interfaceid": "83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.715304] env[63293]: DEBUG nova.network.neutron [req-57ef31f5-ce1c-4958-b96e-482185cba0a2 req-7ab1efbd-f6aa-4f43-830d-4556e046e43d service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updated VIF entry in instance network info cache for port 4c817793-6a1c-4d84-ac42-0487913a257b. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1021.715304] env[63293]: DEBUG nova.network.neutron [req-57ef31f5-ce1c-4958-b96e-482185cba0a2 req-7ab1efbd-f6aa-4f43-830d-4556e046e43d service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updating instance_info_cache with network_info: [{"id": "4c817793-6a1c-4d84-ac42-0487913a257b", "address": "fa:16:3e:18:84:4b", "network": {"id": "8af42ccf-7ab8-424b-a290-8383c292ef85", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1850566934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eec85c5992d64d62a90e290a1aa2d441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c817793-6a", "ovs_interfaceid": "4c817793-6a1c-4d84-ac42-0487913a257b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.812388] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lock "[datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.814136] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Processing image dc0bd04e-b8f4-482d-b243-fa28e2f86738 {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.814136] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738/dc0bd04e-b8f4-482d-b243-fa28e2f86738.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.814136] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired lock "[datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738/dc0bd04e-b8f4-482d-b243-fa28e2f86738.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.814136] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.814136] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c54caa6-a1af-4ba6-911a-66696314dea6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.823645] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.823829] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1021.824599] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8e6c34e-b193-4f8a-ac9c-b4a1d344099b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.828080] env[63293]: DEBUG nova.compute.utils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1021.829438] env[63293]: DEBUG nova.compute.manager [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1021.830097] env[63293]: DEBUG nova.network.neutron [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1021.836837] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1021.836837] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f0e204-d424-ff8c-5108-d048106c23f8" [ 1021.836837] env[63293]: _type = "Task" [ 1021.836837] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.845778] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f0e204-d424-ff8c-5108-d048106c23f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.856699] env[63293]: DEBUG nova.compute.utils [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1021.888241] env[63293]: DEBUG nova.policy [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7a950baa92f4d9ea77e5d6461582e0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec53642df3804e7190615487a426d4a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 1021.947517] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328123, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.023803] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b982de-1b94-4c65-9cab-1047dc5ada93 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.031352] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df174ea8-ba4f-4b70-a0dd-8de38a92bc66 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.061685] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc59f8dd-85be-45e7-a8aa-dc33697fcbb1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.069331] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b9fb48-99b7-4ead-b45d-bc3583eccfaf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.083589] env[63293]: DEBUG nova.compute.provider_tree [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.088524] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Releasing lock "refresh_cache-d605e066-806a-4aac-9af6-f79d542ab2e3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.088524] env[63293]: DEBUG nova.objects.instance [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lazy-loading 'migration_context' on Instance uuid d605e066-806a-4aac-9af6-f79d542ab2e3 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.196304] env[63293]: DEBUG nova.network.neutron [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Successfully created port: a6a90b56-3edd-4270-a9f8-181df8dbb637 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1022.218702] env[63293]: DEBUG oslo_concurrency.lockutils [req-57ef31f5-ce1c-4958-b96e-482185cba0a2 req-7ab1efbd-f6aa-4f43-830d-4556e046e43d service nova] Releasing lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.334405] env[63293]: DEBUG nova.compute.manager [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1022.349024] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Preparing fetch location {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1022.350013] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Fetch image to [datastore1] OSTACK_IMG_96bf553e-927c-4101-bf6b-dcf622b8b0d7/OSTACK_IMG_96bf553e-927c-4101-bf6b-dcf622b8b0d7.vmdk {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1022.350013] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Downloading stream optimized image dc0bd04e-b8f4-482d-b243-fa28e2f86738 to [datastore1] OSTACK_IMG_96bf553e-927c-4101-bf6b-dcf622b8b0d7/OSTACK_IMG_96bf553e-927c-4101-bf6b-dcf622b8b0d7.vmdk on the data store datastore1 as vApp {{(pid=63293) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1022.350013] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Downloading image file data dc0bd04e-b8f4-482d-b243-fa28e2f86738 to the ESX as VM named 'OSTACK_IMG_96bf553e-927c-4101-bf6b-dcf622b8b0d7' {{(pid=63293) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1022.368886] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.013s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.443317] env[63293]: DEBUG oslo_vmware.rw_handles [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1022.443317] env[63293]: value = "resgroup-9" [ 1022.443317] env[63293]: _type = "ResourcePool" [ 1022.443317] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1022.443317] env[63293]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b32a4369-3e8a-4ca7-b0cb-c4ad0ef9853c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.461212] env[63293]: DEBUG oslo_vmware.api [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328123, 'name': PowerOnVM_Task, 'duration_secs': 0.726192} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.462127] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1022.462235] env[63293]: INFO nova.compute.manager [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Took 8.21 seconds to spawn the instance on the hypervisor. [ 1022.462331] env[63293]: DEBUG nova.compute.manager [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.463142] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e72196-35ef-473c-b28f-ab9bd42a1966 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.466924] env[63293]: DEBUG oslo_vmware.rw_handles [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lease: (returnval){ [ 1022.466924] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52832746-fb8a-52ab-1ac6-6310b7dbd461" [ 1022.466924] env[63293]: _type = "HttpNfcLease" [ 1022.466924] env[63293]: } obtained for vApp import into resource pool (val){ [ 1022.466924] env[63293]: value = "resgroup-9" [ 1022.466924] env[63293]: _type = "ResourcePool" [ 1022.466924] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1022.467227] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the lease: (returnval){ [ 1022.467227] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52832746-fb8a-52ab-1ac6-6310b7dbd461" [ 1022.467227] env[63293]: _type = "HttpNfcLease" [ 1022.467227] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1022.479388] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1022.479388] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52832746-fb8a-52ab-1ac6-6310b7dbd461" [ 1022.479388] env[63293]: _type = "HttpNfcLease" [ 1022.479388] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1022.586461] env[63293]: DEBUG nova.scheduler.client.report [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.590334] env[63293]: DEBUG nova.objects.base [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1022.591559] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8021cae-763a-40b4-92b8-ab5825ec012b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.611166] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3f4b232-6ff2-4063-940e-231bf4caa0f2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.616955] env[63293]: DEBUG oslo_vmware.api [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 1022.616955] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5280308c-a09c-ec40-4bed-48e5d7d04b8f" [ 1022.616955] env[63293]: _type = "Task" [ 1022.616955] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.624597] env[63293]: DEBUG oslo_vmware.api [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5280308c-a09c-ec40-4bed-48e5d7d04b8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.980738] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1022.980738] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52832746-fb8a-52ab-1ac6-6310b7dbd461" [ 1022.980738] env[63293]: _type = "HttpNfcLease" [ 1022.980738] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1022.981069] env[63293]: DEBUG oslo_vmware.rw_handles [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1022.981069] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52832746-fb8a-52ab-1ac6-6310b7dbd461" [ 1022.981069] env[63293]: _type = "HttpNfcLease" [ 1022.981069] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1022.981841] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148cfe81-7de1-4931-aa16-2d4116199b89 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.986342] env[63293]: INFO nova.compute.manager [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Took 19.96 seconds to build instance. [ 1022.990811] env[63293]: DEBUG oslo_vmware.rw_handles [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36fc1-d3b5-4a2b-8a64-103dc6735056/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1022.991075] env[63293]: DEBUG oslo_vmware.rw_handles [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36fc1-d3b5-4a2b-8a64-103dc6735056/disk-0.vmdk. {{(pid=63293) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1023.054089] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a8d4f46d-64b5-477b-8914-5ef860222357 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.091987] env[63293]: DEBUG oslo_concurrency.lockutils [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.766s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.094206] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.714s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.096305] env[63293]: INFO nova.compute.claims [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.118117] env[63293]: INFO nova.scheduler.client.report [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Deleted allocations for instance f9ccb216-1435-44c5-ab34-a6388d794551 [ 1023.130928] env[63293]: DEBUG oslo_vmware.api [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5280308c-a09c-ec40-4bed-48e5d7d04b8f, 'name': SearchDatastore_Task, 'duration_secs': 0.007085} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.131354] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.344793] env[63293]: DEBUG nova.compute.manager [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1023.373099] env[63293]: DEBUG nova.virt.hardware [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1023.373529] env[63293]: DEBUG nova.virt.hardware [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1023.373741] env[63293]: DEBUG nova.virt.hardware [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1023.373969] env[63293]: DEBUG nova.virt.hardware [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1023.374207] env[63293]: DEBUG nova.virt.hardware [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1023.374395] env[63293]: DEBUG nova.virt.hardware [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1023.374648] env[63293]: DEBUG nova.virt.hardware [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1023.374956] env[63293]: DEBUG nova.virt.hardware [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1023.375236] env[63293]: DEBUG nova.virt.hardware [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1023.375457] env[63293]: DEBUG nova.virt.hardware [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1023.375789] env[63293]: DEBUG nova.virt.hardware [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.376816] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c8f4f3-0333-4fe9-8d0f-285f232b85b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.393159] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb3997f-ab9c-47fa-9385-23e71d81be90 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.429290] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.429620] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.429905] env[63293]: INFO nova.compute.manager [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Attaching volume 4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6 to /dev/sdb [ 1023.469810] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc15ca1-351c-45e1-974e-217523893971 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.480099] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdbdc0b3-2c9c-4bce-bd2e-0dd5c720e7c9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.488161] env[63293]: DEBUG oslo_concurrency.lockutils [None req-62f1a8c9-2a54-416e-9866-f9c1232f7791 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.472s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.496221] env[63293]: DEBUG nova.virt.block_device [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Updating existing volume attachment record: 07ee4629-eec5-4d5e-8d69-b49ae734b351 {{(pid=63293) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1023.632027] env[63293]: DEBUG oslo_concurrency.lockutils [None req-63472f33-8692-4fab-9ec5-af9d4c7a974f tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "f9ccb216-1435-44c5-ab34-a6388d794551" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.548s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.645010] env[63293]: DEBUG nova.compute.manager [req-97cc4084-9260-47e7-8e88-1814158d97c4 req-05a91994-86d9-41e9-829a-f0e2a9d4cc7d service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Received event network-vif-plugged-a6a90b56-3edd-4270-a9f8-181df8dbb637 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.645259] env[63293]: DEBUG oslo_concurrency.lockutils [req-97cc4084-9260-47e7-8e88-1814158d97c4 req-05a91994-86d9-41e9-829a-f0e2a9d4cc7d service nova] Acquiring lock "97da107a-2e15-4da2-9197-7f85d960ff27-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.645476] env[63293]: DEBUG oslo_concurrency.lockutils [req-97cc4084-9260-47e7-8e88-1814158d97c4 req-05a91994-86d9-41e9-829a-f0e2a9d4cc7d service nova] Lock "97da107a-2e15-4da2-9197-7f85d960ff27-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.645663] env[63293]: DEBUG oslo_concurrency.lockutils [req-97cc4084-9260-47e7-8e88-1814158d97c4 req-05a91994-86d9-41e9-829a-f0e2a9d4cc7d service nova] Lock "97da107a-2e15-4da2-9197-7f85d960ff27-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.645837] env[63293]: DEBUG nova.compute.manager [req-97cc4084-9260-47e7-8e88-1814158d97c4 req-05a91994-86d9-41e9-829a-f0e2a9d4cc7d service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] No waiting events found dispatching network-vif-plugged-a6a90b56-3edd-4270-a9f8-181df8dbb637 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1023.646010] env[63293]: WARNING nova.compute.manager [req-97cc4084-9260-47e7-8e88-1814158d97c4 req-05a91994-86d9-41e9-829a-f0e2a9d4cc7d service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Received unexpected event network-vif-plugged-a6a90b56-3edd-4270-a9f8-181df8dbb637 for instance with vm_state building and task_state spawning. [ 1023.752948] env[63293]: DEBUG nova.network.neutron [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Successfully updated port: a6a90b56-3edd-4270-a9f8-181df8dbb637 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1023.937583] env[63293]: DEBUG nova.compute.manager [req-79c02005-2281-465f-bae4-2a2a09c54719 req-54c69af5-119f-4573-b65f-9799f84202d8 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Received event network-changed-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.937788] env[63293]: DEBUG nova.compute.manager [req-79c02005-2281-465f-bae4-2a2a09c54719 req-54c69af5-119f-4573-b65f-9799f84202d8 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Refreshing instance network info cache due to event network-changed-3313c3a6-17f4-4c97-b19f-9edfd6eb4099. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1023.938055] env[63293]: DEBUG oslo_concurrency.lockutils [req-79c02005-2281-465f-bae4-2a2a09c54719 req-54c69af5-119f-4573-b65f-9799f84202d8 service nova] Acquiring lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.938242] env[63293]: DEBUG oslo_concurrency.lockutils [req-79c02005-2281-465f-bae4-2a2a09c54719 req-54c69af5-119f-4573-b65f-9799f84202d8 service nova] Acquired lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.938568] env[63293]: DEBUG nova.network.neutron [req-79c02005-2281-465f-bae4-2a2a09c54719 req-54c69af5-119f-4573-b65f-9799f84202d8 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Refreshing network info cache for port 3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1024.196275] env[63293]: DEBUG oslo_vmware.rw_handles [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Completed reading data from the image iterator. {{(pid=63293) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1024.196674] env[63293]: DEBUG oslo_vmware.rw_handles [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36fc1-d3b5-4a2b-8a64-103dc6735056/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1024.198239] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe707287-eb00-4cbf-a54a-4a808a818e03 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.208056] env[63293]: DEBUG oslo_vmware.rw_handles [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36fc1-d3b5-4a2b-8a64-103dc6735056/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1024.208056] env[63293]: DEBUG oslo_vmware.rw_handles [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36fc1-d3b5-4a2b-8a64-103dc6735056/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1024.211784] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-f9827672-bc76-4662-b20c-ce27939810db {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.262051] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "refresh_cache-97da107a-2e15-4da2-9197-7f85d960ff27" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.262051] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "refresh_cache-97da107a-2e15-4da2-9197-7f85d960ff27" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.262051] env[63293]: DEBUG nova.network.neutron [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1024.279619] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c4e465-0043-4265-8c2a-d5a3703a02e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.287902] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f4742a-090a-4b11-86d6-cd48f44d7eb8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.321688] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b742795-c568-4653-b7e5-91721bca4375 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.330018] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf4ce3f-39e2-4ae3-be16-4e4173b56cc5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.344186] env[63293]: DEBUG nova.compute.provider_tree [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.409169] env[63293]: DEBUG oslo_vmware.rw_handles [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36fc1-d3b5-4a2b-8a64-103dc6735056/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1024.409554] env[63293]: INFO nova.virt.vmwareapi.images [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Downloaded image file data dc0bd04e-b8f4-482d-b243-fa28e2f86738 [ 1024.410300] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c11807c-428c-49b6-8bb2-732a40863649 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.426143] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e0b82c6-6a7a-456c-9bca-12c9260349be {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.457326] env[63293]: INFO nova.virt.vmwareapi.images [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] The imported VM was unregistered [ 1024.460173] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Caching image {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1024.460173] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Creating directory with path [datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738 {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.460892] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15eae30a-a21d-41be-b364-8f9c93e9f0d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.475681] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Created directory with path [datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738 {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.475939] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_96bf553e-927c-4101-bf6b-dcf622b8b0d7/OSTACK_IMG_96bf553e-927c-4101-bf6b-dcf622b8b0d7.vmdk to [datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738/dc0bd04e-b8f4-482d-b243-fa28e2f86738.vmdk. {{(pid=63293) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1024.478649] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c6e6bd43-4381-45ca-9e47-73508cfb79b9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.487650] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1024.487650] env[63293]: value = "task-1328127" [ 1024.487650] env[63293]: _type = "Task" [ 1024.487650] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.495942] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328127, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.746936] env[63293]: DEBUG nova.network.neutron [req-79c02005-2281-465f-bae4-2a2a09c54719 req-54c69af5-119f-4573-b65f-9799f84202d8 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updated VIF entry in instance network info cache for port 3313c3a6-17f4-4c97-b19f-9edfd6eb4099. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1024.747414] env[63293]: DEBUG nova.network.neutron [req-79c02005-2281-465f-bae4-2a2a09c54719 req-54c69af5-119f-4573-b65f-9799f84202d8 service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updating instance_info_cache with network_info: [{"id": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "address": "fa:16:3e:85:83:c3", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3313c3a6-17", "ovs_interfaceid": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.793359] env[63293]: DEBUG nova.network.neutron [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1024.847014] env[63293]: DEBUG nova.scheduler.client.report [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.939259] env[63293]: DEBUG nova.network.neutron [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Updating instance_info_cache with network_info: [{"id": "a6a90b56-3edd-4270-a9f8-181df8dbb637", "address": "fa:16:3e:fb:03:a7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6a90b56-3e", "ovs_interfaceid": "a6a90b56-3edd-4270-a9f8-181df8dbb637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.000261] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328127, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.002886] env[63293]: DEBUG oslo_concurrency.lockutils [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "35cbf277-b85a-4101-beae-0125fb5a61f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.003164] env[63293]: DEBUG oslo_concurrency.lockutils [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "35cbf277-b85a-4101-beae-0125fb5a61f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.003384] env[63293]: DEBUG oslo_concurrency.lockutils [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "35cbf277-b85a-4101-beae-0125fb5a61f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.003795] env[63293]: DEBUG oslo_concurrency.lockutils [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "35cbf277-b85a-4101-beae-0125fb5a61f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.004013] env[63293]: DEBUG oslo_concurrency.lockutils [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "35cbf277-b85a-4101-beae-0125fb5a61f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.006445] env[63293]: INFO nova.compute.manager [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Terminating instance [ 1025.008685] env[63293]: DEBUG nova.compute.manager [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1025.008928] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1025.009885] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2882208a-02f5-4c65-ab09-14576f2789e7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.018950] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1025.019735] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-105b050c-2ad7-4bda-bbe6-6ed24addf2c8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.027464] env[63293]: DEBUG oslo_vmware.api [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 1025.027464] env[63293]: value = "task-1328128" [ 1025.027464] env[63293]: _type = "Task" [ 1025.027464] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.037361] env[63293]: DEBUG oslo_vmware.api [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328128, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.249883] env[63293]: DEBUG oslo_concurrency.lockutils [req-79c02005-2281-465f-bae4-2a2a09c54719 req-54c69af5-119f-4573-b65f-9799f84202d8 service nova] Releasing lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.352554] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.352942] env[63293]: DEBUG nova.compute.manager [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1025.355476] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.224s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.442451] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "refresh_cache-97da107a-2e15-4da2-9197-7f85d960ff27" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.442858] env[63293]: DEBUG nova.compute.manager [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Instance network_info: |[{"id": "a6a90b56-3edd-4270-a9f8-181df8dbb637", "address": "fa:16:3e:fb:03:a7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6a90b56-3e", "ovs_interfaceid": "a6a90b56-3edd-4270-a9f8-181df8dbb637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1025.443326] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:03:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91c1da19-ab68-4127-bacd-accbaff19651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6a90b56-3edd-4270-a9f8-181df8dbb637', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1025.451262] env[63293]: DEBUG oslo.service.loopingcall [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.451632] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1025.451903] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9b8d042-283e-4179-9e2b-ee3938c91469 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.472528] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.472528] env[63293]: value = "task-1328129" [ 1025.472528] env[63293]: _type = "Task" [ 1025.472528] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.480348] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.496372] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328127, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.538654] env[63293]: DEBUG oslo_vmware.api [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328128, 'name': PowerOffVM_Task, 'duration_secs': 0.213226} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.539068] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.539291] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1025.539565] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66e416e5-982b-44b5-9d77-9bc60a10fa2c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.600799] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1025.601073] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1025.601268] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Deleting the datastore file [datastore2] 35cbf277-b85a-4101-beae-0125fb5a61f5 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.601560] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d4917569-acb6-4a2b-b1b5-16fadf4e78d9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.607867] env[63293]: DEBUG oslo_vmware.api [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for the task: (returnval){ [ 1025.607867] env[63293]: value = "task-1328131" [ 1025.607867] env[63293]: _type = "Task" [ 1025.607867] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.615369] env[63293]: DEBUG oslo_vmware.api [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328131, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.670964] env[63293]: DEBUG nova.compute.manager [req-0f5d36ad-9178-47b9-8e54-678c7ef1006c req-1fc81017-8583-468a-b433-521a9c8b9e52 service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Received event network-changed-a6a90b56-3edd-4270-a9f8-181df8dbb637 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1025.671215] env[63293]: DEBUG nova.compute.manager [req-0f5d36ad-9178-47b9-8e54-678c7ef1006c req-1fc81017-8583-468a-b433-521a9c8b9e52 service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Refreshing instance network info cache due to event network-changed-a6a90b56-3edd-4270-a9f8-181df8dbb637. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1025.671403] env[63293]: DEBUG oslo_concurrency.lockutils [req-0f5d36ad-9178-47b9-8e54-678c7ef1006c req-1fc81017-8583-468a-b433-521a9c8b9e52 service nova] Acquiring lock "refresh_cache-97da107a-2e15-4da2-9197-7f85d960ff27" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.671550] env[63293]: DEBUG oslo_concurrency.lockutils [req-0f5d36ad-9178-47b9-8e54-678c7ef1006c req-1fc81017-8583-468a-b433-521a9c8b9e52 service nova] Acquired lock "refresh_cache-97da107a-2e15-4da2-9197-7f85d960ff27" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.671748] env[63293]: DEBUG nova.network.neutron [req-0f5d36ad-9178-47b9-8e54-678c7ef1006c req-1fc81017-8583-468a-b433-521a9c8b9e52 service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Refreshing network info cache for port a6a90b56-3edd-4270-a9f8-181df8dbb637 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1025.859260] env[63293]: DEBUG nova.compute.utils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1025.860724] env[63293]: DEBUG nova.compute.manager [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1025.860900] env[63293]: DEBUG nova.network.neutron [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1025.920426] env[63293]: DEBUG nova.policy [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b94ec4d2961a43ca9a331a6051fa774d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7744835de0bc4e30b1b29ee5a439cd0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 1025.985275] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.001074] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328127, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.033164] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cc32db-1c7a-4ca7-b2cf-6a715f407754 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.041380] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b09c65ec-7999-4679-a1bf-2c22a9e94a37 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.075100] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26622c9b-abb0-456f-b624-cbe4a6c29322 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.083890] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133fd3d4-9993-4c7f-96a3-db4c13012c59 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.098871] env[63293]: DEBUG nova.compute.provider_tree [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.119080] env[63293]: DEBUG oslo_vmware.api [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Task: {'id': task-1328131, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.311497} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.119400] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.119590] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1026.119773] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1026.120038] env[63293]: INFO nova.compute.manager [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1026.120294] env[63293]: DEBUG oslo.service.loopingcall [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.120526] env[63293]: DEBUG nova.compute.manager [-] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.120597] env[63293]: DEBUG nova.network.neutron [-] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1026.318503] env[63293]: DEBUG nova.network.neutron [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Successfully created port: f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1026.366502] env[63293]: DEBUG nova.compute.manager [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1026.392435] env[63293]: DEBUG nova.network.neutron [req-0f5d36ad-9178-47b9-8e54-678c7ef1006c req-1fc81017-8583-468a-b433-521a9c8b9e52 service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Updated VIF entry in instance network info cache for port a6a90b56-3edd-4270-a9f8-181df8dbb637. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1026.392823] env[63293]: DEBUG nova.network.neutron [req-0f5d36ad-9178-47b9-8e54-678c7ef1006c req-1fc81017-8583-468a-b433-521a9c8b9e52 service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Updating instance_info_cache with network_info: [{"id": "a6a90b56-3edd-4270-a9f8-181df8dbb637", "address": "fa:16:3e:fb:03:a7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6a90b56-3e", "ovs_interfaceid": "a6a90b56-3edd-4270-a9f8-181df8dbb637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.484388] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.503136] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328127, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.603132] env[63293]: DEBUG nova.scheduler.client.report [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.895885] env[63293]: DEBUG oslo_concurrency.lockutils [req-0f5d36ad-9178-47b9-8e54-678c7ef1006c req-1fc81017-8583-468a-b433-521a9c8b9e52 service nova] Releasing lock "refresh_cache-97da107a-2e15-4da2-9197-7f85d960ff27" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.985582] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.999876] env[63293]: DEBUG nova.network.neutron [-] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.001349] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328127, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.377793] env[63293]: DEBUG nova.compute.manager [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1027.404626] env[63293]: DEBUG nova.virt.hardware [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1027.404888] env[63293]: DEBUG nova.virt.hardware [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1027.405084] env[63293]: DEBUG nova.virt.hardware [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1027.405284] env[63293]: DEBUG nova.virt.hardware [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1027.405516] env[63293]: DEBUG nova.virt.hardware [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1027.405700] env[63293]: DEBUG nova.virt.hardware [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1027.405939] env[63293]: DEBUG nova.virt.hardware [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1027.406126] env[63293]: DEBUG nova.virt.hardware [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1027.406393] env[63293]: DEBUG nova.virt.hardware [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1027.406569] env[63293]: DEBUG nova.virt.hardware [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1027.406853] env[63293]: DEBUG nova.virt.hardware [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.407843] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36acb97-0034-4346-b48d-380594e2b0d6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.416972] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937dc300-44c3-4041-9346-9c02ce9264c3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.483635] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.500395] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328127, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.502988] env[63293]: INFO nova.compute.manager [-] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Took 1.38 seconds to deallocate network for instance. [ 1027.613886] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.258s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.696783] env[63293]: DEBUG nova.compute.manager [req-4fe61c65-c239-4410-a722-6353ac18b9e8 req-33845e96-02af-4e36-abfb-8b6c91781abf service nova] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Received event network-vif-deleted-df1c83c6-23d7-439b-a496-234c48738d81 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1027.985326] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.000219] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328127, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.009348] env[63293]: DEBUG oslo_concurrency.lockutils [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.009600] env[63293]: DEBUG oslo_concurrency.lockutils [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.009954] env[63293]: DEBUG nova.objects.instance [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lazy-loading 'resources' on Instance uuid 35cbf277-b85a-4101-beae-0125fb5a61f5 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.049383] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Volume attach. Driver type: vmdk {{(pid=63293) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1028.049633] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283832', 'volume_id': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'name': 'volume-4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9937a3c4-c3a6-423b-ac4f-50b7ec5879d2', 'attached_at': '', 'detached_at': '', 'volume_id': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'serial': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1028.050631] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31867e1b-f4c0-470d-941f-f4120ef126fb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.068488] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4362edd4-66a6-4cf3-aa8b-b643c1872e4a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.094141] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] volume-4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6/volume-4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.094602] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f90de2d6-54fc-4f66-9a85-47aabcfb31ab {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.107609] env[63293]: DEBUG nova.network.neutron [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Successfully updated port: f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.114279] env[63293]: DEBUG oslo_vmware.api [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 1028.114279] env[63293]: value = "task-1328133" [ 1028.114279] env[63293]: _type = "Task" [ 1028.114279] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.123452] env[63293]: DEBUG oslo_vmware.api [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328133, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.178073] env[63293]: INFO nova.scheduler.client.report [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleted allocation for migration 42d73401-baaf-4819-bf86-bb2fc639388e [ 1028.484660] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.500567] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328127, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.678216} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.500920] env[63293]: INFO nova.virt.vmwareapi.ds_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_96bf553e-927c-4101-bf6b-dcf622b8b0d7/OSTACK_IMG_96bf553e-927c-4101-bf6b-dcf622b8b0d7.vmdk to [datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738/dc0bd04e-b8f4-482d-b243-fa28e2f86738.vmdk. [ 1028.501199] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Cleaning up location [datastore1] OSTACK_IMG_96bf553e-927c-4101-bf6b-dcf622b8b0d7 {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1028.501463] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_96bf553e-927c-4101-bf6b-dcf622b8b0d7 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.501751] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-207c0d3a-ff4f-4de6-88a0-1c1f830e0960 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.508928] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1028.508928] env[63293]: value = "task-1328134" [ 1028.508928] env[63293]: _type = "Task" [ 1028.508928] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.518479] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328134, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.612125] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "refresh_cache-46a9329b-6e3b-454f-8a50-30ff7bc719f0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.612275] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "refresh_cache-46a9329b-6e3b-454f-8a50-30ff7bc719f0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.612446] env[63293]: DEBUG nova.network.neutron [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.618336] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48bb6537-47c5-4087-ae7f-30fe8b7dacc5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.626933] env[63293]: DEBUG oslo_vmware.api [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328133, 'name': ReconfigVM_Task, 'duration_secs': 0.338371} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.628735] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Reconfigured VM instance instance-0000005f to attach disk [datastore1] volume-4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6/volume-4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.633526] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-804dbceb-edda-4b34-a990-9413f373ab9a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.643543] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe27850-a1aa-4b53-a773-c99e78ea6b29 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.674880] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6f88f5-5320-4fb3-836e-086ec90f6cbc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.677243] env[63293]: DEBUG oslo_vmware.api [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 1028.677243] env[63293]: value = "task-1328135" [ 1028.677243] env[63293]: _type = "Task" [ 1028.677243] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.683363] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9fb71cd-48b3-49f1-9357-b5a9682bd09c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.691028] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.245s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.691912] env[63293]: DEBUG oslo_vmware.api [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.702100] env[63293]: DEBUG nova.compute.provider_tree [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.985780] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.018573] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328134, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201755} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.018854] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.019039] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lock "[datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738/dc0bd04e-b8f4-482d-b243-fa28e2f86738.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.019290] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738/dc0bd04e-b8f4-482d-b243-fa28e2f86738.vmdk to [datastore1] f429869f-9e9f-443e-a980-67a80c3bd799/f429869f-9e9f-443e-a980-67a80c3bd799.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1029.019540] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4b88f390-e854-4489-baf5-b4582a42820b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.026302] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1029.026302] env[63293]: value = "task-1328136" [ 1029.026302] env[63293]: _type = "Task" [ 1029.026302] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.033751] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328136, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.157292] env[63293]: DEBUG nova.network.neutron [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1029.188019] env[63293]: DEBUG oslo_vmware.api [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328135, 'name': ReconfigVM_Task, 'duration_secs': 0.135711} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.188200] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283832', 'volume_id': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'name': 'volume-4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9937a3c4-c3a6-423b-ac4f-50b7ec5879d2', 'attached_at': '', 'detached_at': '', 'volume_id': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'serial': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1029.206143] env[63293]: DEBUG nova.scheduler.client.report [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.298216] env[63293]: DEBUG nova.network.neutron [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Updating instance_info_cache with network_info: [{"id": "f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b", "address": "fa:16:3e:26:25:e9", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4c5ccaf-1c", "ovs_interfaceid": "f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.487927] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.536686] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328136, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.712080] env[63293]: DEBUG oslo_concurrency.lockutils [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.702s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.735127] env[63293]: DEBUG nova.compute.manager [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Received event network-vif-plugged-f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.735127] env[63293]: DEBUG oslo_concurrency.lockutils [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] Acquiring lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.735328] env[63293]: DEBUG oslo_concurrency.lockutils [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] Lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.735429] env[63293]: DEBUG oslo_concurrency.lockutils [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] Lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.735888] env[63293]: DEBUG nova.compute.manager [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] No waiting events found dispatching network-vif-plugged-f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1029.736138] env[63293]: WARNING nova.compute.manager [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Received unexpected event network-vif-plugged-f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b for instance with vm_state building and task_state spawning. [ 1029.736312] env[63293]: DEBUG nova.compute.manager [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Received event network-changed-f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.736512] env[63293]: DEBUG nova.compute.manager [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Refreshing instance network info cache due to event network-changed-f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1029.736707] env[63293]: DEBUG oslo_concurrency.lockutils [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] Acquiring lock "refresh_cache-46a9329b-6e3b-454f-8a50-30ff7bc719f0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.737703] env[63293]: INFO nova.scheduler.client.report [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Deleted allocations for instance 35cbf277-b85a-4101-beae-0125fb5a61f5 [ 1029.800683] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "refresh_cache-46a9329b-6e3b-454f-8a50-30ff7bc719f0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.801064] env[63293]: DEBUG nova.compute.manager [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Instance network_info: |[{"id": "f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b", "address": "fa:16:3e:26:25:e9", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4c5ccaf-1c", "ovs_interfaceid": "f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1029.801397] env[63293]: DEBUG oslo_concurrency.lockutils [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] Acquired lock "refresh_cache-46a9329b-6e3b-454f-8a50-30ff7bc719f0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.801616] env[63293]: DEBUG nova.network.neutron [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Refreshing network info cache for port f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1029.802848] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:25:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ef746c57-cd18-4883-a0e9-c52937aaf41d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1029.811621] env[63293]: DEBUG oslo.service.loopingcall [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.812885] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1029.813175] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3f5ac3b-987e-4151-837c-41aee89ff20e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.836545] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.836545] env[63293]: value = "task-1328137" [ 1029.836545] env[63293]: _type = "Task" [ 1029.836545] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.845440] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328137, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.969250] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "d605e066-806a-4aac-9af6-f79d542ab2e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.969542] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.969812] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "d605e066-806a-4aac-9af6-f79d542ab2e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.970062] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.970276] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.972953] env[63293]: INFO nova.compute.manager [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Terminating instance [ 1029.975140] env[63293]: DEBUG nova.compute.manager [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1029.975409] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1029.976408] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5566f339-2df6-4eda-8e9a-23206d86db74 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.987931] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1029.991663] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f804c254-8a8a-4d3b-a52a-acf3b00b09de {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.993315] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.998551] env[63293]: DEBUG oslo_vmware.api [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 1029.998551] env[63293]: value = "task-1328138" [ 1029.998551] env[63293]: _type = "Task" [ 1029.998551] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.007061] env[63293]: DEBUG oslo_vmware.api [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328138, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.036315] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328136, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.229343] env[63293]: DEBUG nova.objects.instance [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'flavor' on Instance uuid 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.245486] env[63293]: DEBUG oslo_concurrency.lockutils [None req-efee1044-3b35-4d09-99d3-6cf26d90f91c tempest-AttachInterfacesTestJSON-371461994 tempest-AttachInterfacesTestJSON-371461994-project-member] Lock "35cbf277-b85a-4101-beae-0125fb5a61f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.242s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.347497] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328137, 'name': CreateVM_Task, 'duration_secs': 0.356219} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.350212] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1030.351102] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.351225] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.351571] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1030.352395] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47bf0104-8f5e-4e62-ab7a-4af41774f52f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.357637] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1030.357637] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527e808c-5e63-feaf-0531-8727b08f2621" [ 1030.357637] env[63293]: _type = "Task" [ 1030.357637] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.367044] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527e808c-5e63-feaf-0531-8727b08f2621, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.493489] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.507273] env[63293]: DEBUG oslo_vmware.api [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328138, 'name': PowerOffVM_Task, 'duration_secs': 0.394809} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.507585] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1030.507763] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1030.508035] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3087503-ae03-4fb2-9ff5-68924c7cd250 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.537891] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328136, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.609684] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1030.609684] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1030.609684] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleting the datastore file [datastore1] d605e066-806a-4aac-9af6-f79d542ab2e3 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.610606] env[63293]: DEBUG nova.network.neutron [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Updated VIF entry in instance network info cache for port f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1030.611055] env[63293]: DEBUG nova.network.neutron [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Updating instance_info_cache with network_info: [{"id": "f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b", "address": "fa:16:3e:26:25:e9", "network": {"id": "80025d23-bc97-4ed5-bece-102499484f3e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1209303649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7744835de0bc4e30b1b29ee5a439cd0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4c5ccaf-1c", "ovs_interfaceid": "f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.612479] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2322088f-3220-4863-a37d-0535f5ce37e0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.619887] env[63293]: DEBUG oslo_vmware.api [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for the task: (returnval){ [ 1030.619887] env[63293]: value = "task-1328140" [ 1030.619887] env[63293]: _type = "Task" [ 1030.619887] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.631316] env[63293]: DEBUG oslo_vmware.api [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.735288] env[63293]: DEBUG oslo_concurrency.lockutils [None req-68d24bd0-9475-4fa0-8437-b044591092f4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.306s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.868984] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]527e808c-5e63-feaf-0531-8727b08f2621, 'name': SearchDatastore_Task, 'duration_secs': 0.015034} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.869375] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.869622] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1030.869874] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.870043] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.870235] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1030.870519] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1cb88145-104e-45f5-93a8-fba0695e0257 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.888610] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1030.888610] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1030.889313] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abafd793-d25c-4301-94c8-7f7771ad0fdb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.895163] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1030.895163] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f67872-e71f-4f6d-e7a3-ebbff8b00996" [ 1030.895163] env[63293]: _type = "Task" [ 1030.895163] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.902918] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f67872-e71f-4f6d-e7a3-ebbff8b00996, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.990313] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328129, 'name': CreateVM_Task, 'duration_secs': 5.505493} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.990502] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1030.991273] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.991448] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.991831] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1030.992121] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d108651-714f-4108-81e4-a553538d4f5e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.997206] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1030.997206] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52214c83-aed5-e1b4-179e-fb041b37cd5d" [ 1030.997206] env[63293]: _type = "Task" [ 1030.997206] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.005989] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52214c83-aed5-e1b4-179e-fb041b37cd5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.037151] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328136, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.116120] env[63293]: DEBUG oslo_concurrency.lockutils [req-65ffa7cf-aea2-4a63-849a-729727096f3c req-4a6d2197-c032-406a-b6c1-12cc4d4b6553 service nova] Releasing lock "refresh_cache-46a9329b-6e3b-454f-8a50-30ff7bc719f0" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.131745] env[63293]: DEBUG oslo_vmware.api [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.242071] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1031.242322] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1031.242469] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Starting heal instance info cache {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1031.410064] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f67872-e71f-4f6d-e7a3-ebbff8b00996, 'name': SearchDatastore_Task, 'duration_secs': 0.012} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.410064] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38c7f754-7cb5-4a3e-aa2c-3f9217c682e6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.413983] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1031.413983] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524fc33a-9263-be36-09c4-abf20650c740" [ 1031.413983] env[63293]: _type = "Task" [ 1031.413983] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.422276] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524fc33a-9263-be36-09c4-abf20650c740, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.507620] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52214c83-aed5-e1b4-179e-fb041b37cd5d, 'name': SearchDatastore_Task, 'duration_secs': 0.084843} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.507931] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.508274] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1031.508434] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.508586] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.508770] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.509499] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6eddf63d-b15b-4223-81ec-b7dc7f46a385 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.525212] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.525363] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1031.526170] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5611a40e-1a44-4981-92d0-332447fdb691 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.540369] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1031.540369] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5237c272-8a03-e022-9d93-914a28cc28f8" [ 1031.540369] env[63293]: _type = "Task" [ 1031.540369] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.540726] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328136, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.551063] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5237c272-8a03-e022-9d93-914a28cc28f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.630602] env[63293]: DEBUG oslo_vmware.api [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.640845] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.641166] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.925215] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]524fc33a-9263-be36-09c4-abf20650c740, 'name': SearchDatastore_Task, 'duration_secs': 0.019898} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.925549] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.925839] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 46a9329b-6e3b-454f-8a50-30ff7bc719f0/46a9329b-6e3b-454f-8a50-30ff7bc719f0.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1031.926180] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7807b1b2-f82d-4c0b-b4e2-a36063edf8e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.935058] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1031.935058] env[63293]: value = "task-1328141" [ 1031.935058] env[63293]: _type = "Task" [ 1031.935058] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.943418] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328141, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.039404] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328136, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.669642} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.039705] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/dc0bd04e-b8f4-482d-b243-fa28e2f86738/dc0bd04e-b8f4-482d-b243-fa28e2f86738.vmdk to [datastore1] f429869f-9e9f-443e-a980-67a80c3bd799/f429869f-9e9f-443e-a980-67a80c3bd799.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1032.040566] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c31a7e0c-3be9-4fc9-ae43-5adef2a14f4c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.051377] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5237c272-8a03-e022-9d93-914a28cc28f8, 'name': SearchDatastore_Task, 'duration_secs': 0.067259} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.069332] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] f429869f-9e9f-443e-a980-67a80c3bd799/f429869f-9e9f-443e-a980-67a80c3bd799.vmdk or device None with type streamOptimized {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.069651] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee7d15a2-ac88-408f-b9c5-39e5f6479f93 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.072265] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56d1c0c4-e034-45d7-8f0f-b73c8a92dc14 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.089903] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1032.089903] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5238a04f-1726-5064-322d-7478d248ce4d" [ 1032.089903] env[63293]: _type = "Task" [ 1032.089903] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.091379] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1032.091379] env[63293]: value = "task-1328142" [ 1032.091379] env[63293]: _type = "Task" [ 1032.091379] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.102496] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5238a04f-1726-5064-322d-7478d248ce4d, 'name': SearchDatastore_Task} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.105711] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.105991] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 97da107a-2e15-4da2-9197-7f85d960ff27/97da107a-2e15-4da2-9197-7f85d960ff27.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1032.106329] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328142, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.106551] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ce00465-bce6-4090-8ee6-28db39677cd6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.117291] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1032.117291] env[63293]: value = "task-1328143" [ 1032.117291] env[63293]: _type = "Task" [ 1032.117291] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.129922] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328143, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.135228] env[63293]: DEBUG oslo_vmware.api [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Task: {'id': task-1328140, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.362666} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.135509] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1032.135765] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1032.136058] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1032.136224] env[63293]: INFO nova.compute.manager [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Took 2.16 seconds to destroy the instance on the hypervisor. [ 1032.136479] env[63293]: DEBUG oslo.service.loopingcall [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.136737] env[63293]: DEBUG nova.compute.manager [-] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1032.136835] env[63293]: DEBUG nova.network.neutron [-] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1032.144137] env[63293]: DEBUG nova.compute.utils [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1032.450881] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328141, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47358} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.451124] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 46a9329b-6e3b-454f-8a50-30ff7bc719f0/46a9329b-6e3b-454f-8a50-30ff7bc719f0.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1032.451358] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.451676] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21999daa-4fbf-4982-bd68-76cba300da69 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.459487] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1032.459487] env[63293]: value = "task-1328144" [ 1032.459487] env[63293]: _type = "Task" [ 1032.459487] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.470341] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328144, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.581985] env[63293]: DEBUG nova.compute.manager [req-24d948bf-76dd-4540-a7af-0fe3375dd047 req-fd72933c-53bb-4633-be4a-5c15159b42d0 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Received event network-vif-deleted-83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.582297] env[63293]: INFO nova.compute.manager [req-24d948bf-76dd-4540-a7af-0fe3375dd047 req-fd72933c-53bb-4633-be4a-5c15159b42d0 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Neutron deleted interface 83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970; detaching it from the instance and deleting it from the info cache [ 1032.582424] env[63293]: DEBUG nova.network.neutron [req-24d948bf-76dd-4540-a7af-0fe3375dd047 req-fd72933c-53bb-4633-be4a-5c15159b42d0 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.608017] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328142, 'name': ReconfigVM_Task, 'duration_secs': 0.314253} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.609028] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Reconfigured VM instance instance-00000055 to attach disk [datastore1] f429869f-9e9f-443e-a980-67a80c3bd799/f429869f-9e9f-443e-a980-67a80c3bd799.vmdk or device None with type streamOptimized {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.609801] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-81465608-9ae3-475e-baca-215a17a99a07 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.617719] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1032.617719] env[63293]: value = "task-1328145" [ 1032.617719] env[63293]: _type = "Task" [ 1032.617719] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.632028] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328145, 'name': Rename_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.633251] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328143, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.647905] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.969336] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328144, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112726} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.969621] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.970400] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255474b8-79ef-4b91-8469-54ecb51b9fac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.992114] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 46a9329b-6e3b-454f-8a50-30ff7bc719f0/46a9329b-6e3b-454f-8a50-30ff7bc719f0.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.992393] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86ad0408-81f0-44ef-b04b-3dc0600de18b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.012242] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1033.012242] env[63293]: value = "task-1328146" [ 1033.012242] env[63293]: _type = "Task" [ 1033.012242] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.012572] env[63293]: DEBUG nova.network.neutron [-] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.021594] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328146, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.085324] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d290d36e-b2c3-4df9-a019-0e1b1abdf4c1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.095196] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af73cf24-35ce-40d3-b5b9-323033d239bb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.124136] env[63293]: DEBUG nova.compute.manager [req-24d948bf-76dd-4540-a7af-0fe3375dd047 req-fd72933c-53bb-4633-be4a-5c15159b42d0 service nova] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Detach interface failed, port_id=83a30fb0-a4d3-46e3-9d97-1d1eb7d4a970, reason: Instance d605e066-806a-4aac-9af6-f79d542ab2e3 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1033.134889] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328143, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.687975} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.137922] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 97da107a-2e15-4da2-9197-7f85d960ff27/97da107a-2e15-4da2-9197-7f85d960ff27.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1033.138180] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1033.138460] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328145, 'name': Rename_Task, 'duration_secs': 0.194499} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.138748] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef7e44a8-db75-44c3-ab36-1217cd98492b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.140605] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1033.140887] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6502fcc1-48bc-4348-bacc-8b82a6871134 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.151231] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1033.151231] env[63293]: value = "task-1328147" [ 1033.151231] env[63293]: _type = "Task" [ 1033.151231] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.153401] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1033.153401] env[63293]: value = "task-1328148" [ 1033.153401] env[63293]: _type = "Task" [ 1033.153401] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.166610] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328148, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.169570] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328147, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.517238] env[63293]: INFO nova.compute.manager [-] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Took 1.38 seconds to deallocate network for instance. [ 1033.526568] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328146, 'name': ReconfigVM_Task, 'duration_secs': 0.355865} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.526568] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 46a9329b-6e3b-454f-8a50-30ff7bc719f0/46a9329b-6e3b-454f-8a50-30ff7bc719f0.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.527232] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-89e94f38-d83e-471c-ab4b-9604dd23015a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.533875] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1033.533875] env[63293]: value = "task-1328149" [ 1033.533875] env[63293]: _type = "Task" [ 1033.533875] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.542681] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328149, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.669208] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328147, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.671791] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328148, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.720135] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.720135] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.720135] env[63293]: INFO nova.compute.manager [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Attaching volume 7a0b3a44-ef09-4b4d-9007-6d4753db1906 to /dev/sdc [ 1033.756190] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba19671-bdfd-4313-97a2-1c3769592111 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.766376] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63697a70-849c-44d0-84ed-9c7cec60d63b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.781942] env[63293]: DEBUG nova.virt.block_device [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Updating existing volume attachment record: e17b097a-a2e5-4aec-a88a-b4caa03a381d {{(pid=63293) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1034.028993] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.029307] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.029547] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.045031] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328149, 'name': Rename_Task, 'duration_secs': 0.292932} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.045205] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.045477] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-efd1c93c-38f5-41ce-aee0-a509e9486c2d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.053020] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1034.053020] env[63293]: value = "task-1328151" [ 1034.053020] env[63293]: _type = "Task" [ 1034.053020] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.057434] env[63293]: INFO nova.scheduler.client.report [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Deleted allocations for instance d605e066-806a-4aac-9af6-f79d542ab2e3 [ 1034.064891] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328151, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.162745] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328147, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.172588] env[63293]: DEBUG oslo_vmware.api [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328148, 'name': PowerOnVM_Task, 'duration_secs': 0.932289} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.173286] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.261983] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Didn't find any instances for network info cache update. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1034.262316] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.262499] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.262562] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.262704] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.262844] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.262986] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.263138] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63293) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1034.263389] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.283621] env[63293]: DEBUG nova.compute.manager [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.284555] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecbfbd6c-e1b6-4f17-9c50-ab8a2d348e26 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.563666] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328151, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.569101] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11541104-b017-4dc6-92e6-07b5305e0b97 tempest-DeleteServersTestJSON-870834261 tempest-DeleteServersTestJSON-870834261-project-member] Lock "d605e066-806a-4aac-9af6-f79d542ab2e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.599s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.663381] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328147, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.083134} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.664433] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1034.665313] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50f588d-cd4a-499d-8497-2a6d97c31db8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.692985] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 97da107a-2e15-4da2-9197-7f85d960ff27/97da107a-2e15-4da2-9197-7f85d960ff27.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.693456] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-964d9353-702f-46aa-8b13-3f8719da24c5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.719202] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1034.719202] env[63293]: value = "task-1328152" [ 1034.719202] env[63293]: _type = "Task" [ 1034.719202] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.730672] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328152, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.768987] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.769274] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.769467] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.769647] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63293) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1034.770623] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f12e7553-31fc-4ecc-9252-16ed3b5f4b98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.778778] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6427e7c-3b6a-41c7-af9b-fe395d4e57a7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.799243] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255f1913-a185-4b0b-8304-55a717698984 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.804973] env[63293]: DEBUG oslo_concurrency.lockutils [None req-33d69089-60ee-4869-84e0-12bd010be339 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "f429869f-9e9f-443e-a980-67a80c3bd799" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 24.382s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.808498] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1fc5c79-ca93-48b5-a943-c17db9488fdf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.845920] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179899MB free_disk=132GB free_vcpus=48 pci_devices=None {{(pid=63293) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1034.846376] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.846993] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.065808] env[63293]: DEBUG oslo_vmware.api [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328151, 'name': PowerOnVM_Task, 'duration_secs': 0.81848} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.066281] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1035.066609] env[63293]: INFO nova.compute.manager [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Took 7.69 seconds to spawn the instance on the hypervisor. [ 1035.066722] env[63293]: DEBUG nova.compute.manager [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1035.067504] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee189cfe-5bf7-4ca7-9427-7dd0498caeec {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.229707] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328152, 'name': ReconfigVM_Task, 'duration_secs': 0.328378} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.229997] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 97da107a-2e15-4da2-9197-7f85d960ff27/97da107a-2e15-4da2-9197-7f85d960ff27.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.230643] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4151b12-1bab-4a26-abaa-262ab26cffd2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.237318] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1035.237318] env[63293]: value = "task-1328153" [ 1035.237318] env[63293]: _type = "Task" [ 1035.237318] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.245849] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328153, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.588991] env[63293]: INFO nova.compute.manager [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Took 19.22 seconds to build instance. [ 1035.746726] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328153, 'name': Rename_Task, 'duration_secs': 0.161933} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.747064] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.747341] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d6225ca-1d9c-4f67-a355-125c8d9e8793 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.756202] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1035.756202] env[63293]: value = "task-1328154" [ 1035.756202] env[63293]: _type = "Task" [ 1035.756202] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.763981] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328154, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.887342] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1035.887504] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1035.887630] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance bea1ea16-c224-4a0e-86ed-c2aaf920d3e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1035.887879] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1035.887879] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance f429869f-9e9f-443e-a980-67a80c3bd799 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1035.888010] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 97da107a-2e15-4da2-9197-7f85d960ff27 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1035.888125] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 46a9329b-6e3b-454f-8a50-30ff7bc719f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1035.888319] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1035.888459] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1035.962146] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540a7508-ce9e-4d2d-8632-5a28322b990c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.969196] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-feebdea5-48f4-45b6-8797-0b83dd32bcba tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Suspending the VM {{(pid=63293) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1035.973223] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-d9395581-25c0-426c-a1e0-fe04e14b3352 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.981319] env[63293]: DEBUG oslo_vmware.api [None req-feebdea5-48f4-45b6-8797-0b83dd32bcba tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1035.981319] env[63293]: value = "task-1328155" [ 1035.981319] env[63293]: _type = "Task" [ 1035.981319] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.992038] env[63293]: DEBUG oslo_vmware.api [None req-feebdea5-48f4-45b6-8797-0b83dd32bcba tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328155, 'name': SuspendVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.027349] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f94abf6-7ed1-4536-98af-c099c97f47a9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.035714] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506f7165-5974-4b23-b046-151980ff0457 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.073116] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a9c24cc-06a2-4953-bcf7-e3f072261224 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.080896] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec221d1-eb99-4eee-bbf5-db2b3df2c856 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.095813] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f73887e4-f4ad-496b-9c3b-2d6e1923c8b3 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.736s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.096315] env[63293]: DEBUG nova.compute.provider_tree [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.270582] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328154, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.496150] env[63293]: DEBUG oslo_vmware.api [None req-feebdea5-48f4-45b6-8797-0b83dd32bcba tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328155, 'name': SuspendVM_Task} progress is 58%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.604672] env[63293]: DEBUG nova.scheduler.client.report [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.771284] env[63293]: DEBUG oslo_vmware.api [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328154, 'name': PowerOnVM_Task, 'duration_secs': 0.769099} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.772213] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.772213] env[63293]: INFO nova.compute.manager [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Took 13.43 seconds to spawn the instance on the hypervisor. [ 1036.772599] env[63293]: DEBUG nova.compute.manager [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.773582] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b4474c0-2315-42de-91e2-efd0a1f8f0cf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.809848] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "07418025-09fc-4793-a8ea-f059f42251ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.810100] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "07418025-09fc-4793-a8ea-f059f42251ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.871289] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.871570] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.871819] env[63293]: DEBUG nova.compute.manager [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.872753] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4320709f-7fa1-40fd-9477-9563b3b76fa0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.880246] env[63293]: DEBUG nova.compute.manager [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63293) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1036.880839] env[63293]: DEBUG nova.objects.instance [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lazy-loading 'flavor' on Instance uuid 46a9329b-6e3b-454f-8a50-30ff7bc719f0 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.992812] env[63293]: DEBUG oslo_vmware.api [None req-feebdea5-48f4-45b6-8797-0b83dd32bcba tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328155, 'name': SuspendVM_Task, 'duration_secs': 0.840391} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.993134] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-feebdea5-48f4-45b6-8797-0b83dd32bcba tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Suspended the VM {{(pid=63293) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1036.993325] env[63293]: DEBUG nova.compute.manager [None req-feebdea5-48f4-45b6-8797-0b83dd32bcba tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.994136] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7731f5a8-f281-4623-b257-b8cbc62e004d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.108407] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1037.108751] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.262s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.296945] env[63293]: INFO nova.compute.manager [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Took 24.48 seconds to build instance. [ 1037.312222] env[63293]: DEBUG nova.compute.manager [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1037.391398] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.391751] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f77e951b-2628-4c25-ab79-b78388b0deda {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.400253] env[63293]: DEBUG oslo_vmware.api [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1037.400253] env[63293]: value = "task-1328157" [ 1037.400253] env[63293]: _type = "Task" [ 1037.400253] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.412154] env[63293]: DEBUG oslo_vmware.api [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328157, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.624382] env[63293]: DEBUG nova.compute.manager [req-cf5f0b6a-8271-46f5-a196-58dce84f9ec6 req-7ee54622-83eb-474f-a6ca-26a7ebe2d117 service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Received event network-changed-a6a90b56-3edd-4270-a9f8-181df8dbb637 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.624587] env[63293]: DEBUG nova.compute.manager [req-cf5f0b6a-8271-46f5-a196-58dce84f9ec6 req-7ee54622-83eb-474f-a6ca-26a7ebe2d117 service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Refreshing instance network info cache due to event network-changed-a6a90b56-3edd-4270-a9f8-181df8dbb637. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1037.624842] env[63293]: DEBUG oslo_concurrency.lockutils [req-cf5f0b6a-8271-46f5-a196-58dce84f9ec6 req-7ee54622-83eb-474f-a6ca-26a7ebe2d117 service nova] Acquiring lock "refresh_cache-97da107a-2e15-4da2-9197-7f85d960ff27" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.624951] env[63293]: DEBUG oslo_concurrency.lockutils [req-cf5f0b6a-8271-46f5-a196-58dce84f9ec6 req-7ee54622-83eb-474f-a6ca-26a7ebe2d117 service nova] Acquired lock "refresh_cache-97da107a-2e15-4da2-9197-7f85d960ff27" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.625194] env[63293]: DEBUG nova.network.neutron [req-cf5f0b6a-8271-46f5-a196-58dce84f9ec6 req-7ee54622-83eb-474f-a6ca-26a7ebe2d117 service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Refreshing network info cache for port a6a90b56-3edd-4270-a9f8-181df8dbb637 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1037.798768] env[63293]: DEBUG oslo_concurrency.lockutils [None req-280094ef-20b2-4dd9-94b9-5c959d93bc61 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.999s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.840128] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.841103] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.842737] env[63293]: INFO nova.compute.claims [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1037.910485] env[63293]: DEBUG oslo_vmware.api [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328157, 'name': PowerOffVM_Task, 'duration_secs': 0.185677} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.911029] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.911374] env[63293]: DEBUG nova.compute.manager [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1037.912317] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9eaeb8-7d34-4a06-b197-076eeda70ed0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.331263] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Volume attach. Driver type: vmdk {{(pid=63293) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1038.332769] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283835', 'volume_id': '7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'name': 'volume-7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9937a3c4-c3a6-423b-ac4f-50b7ec5879d2', 'attached_at': '', 'detached_at': '', 'volume_id': '7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'serial': '7a0b3a44-ef09-4b4d-9007-6d4753db1906'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1038.332769] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4434e56-ae74-4c39-a518-9088e3b908e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.353254] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b3f471-36ae-4b9e-8ec8-a271c248050a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.379103] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] volume-7a0b3a44-ef09-4b4d-9007-6d4753db1906/volume-7a0b3a44-ef09-4b4d-9007-6d4753db1906.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.380705] env[63293]: DEBUG nova.network.neutron [req-cf5f0b6a-8271-46f5-a196-58dce84f9ec6 req-7ee54622-83eb-474f-a6ca-26a7ebe2d117 service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Updated VIF entry in instance network info cache for port a6a90b56-3edd-4270-a9f8-181df8dbb637. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1038.381042] env[63293]: DEBUG nova.network.neutron [req-cf5f0b6a-8271-46f5-a196-58dce84f9ec6 req-7ee54622-83eb-474f-a6ca-26a7ebe2d117 service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Updating instance_info_cache with network_info: [{"id": "a6a90b56-3edd-4270-a9f8-181df8dbb637", "address": "fa:16:3e:fb:03:a7", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6a90b56-3e", "ovs_interfaceid": "a6a90b56-3edd-4270-a9f8-181df8dbb637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.382206] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b19bafa-7b95-4d8a-8466-24499e25a0ec {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.403258] env[63293]: DEBUG oslo_vmware.api [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 1038.403258] env[63293]: value = "task-1328158" [ 1038.403258] env[63293]: _type = "Task" [ 1038.403258] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.412279] env[63293]: DEBUG oslo_vmware.api [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328158, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.427715] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c8493606-2d9c-443b-9ac2-2da02037a80c tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.556s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.433264] env[63293]: INFO nova.compute.manager [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Resuming [ 1038.433819] env[63293]: DEBUG nova.objects.instance [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lazy-loading 'flavor' on Instance uuid f429869f-9e9f-443e-a980-67a80c3bd799 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.875305] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "0041acd3-de76-4745-a1ab-c6563f774f33" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.876325] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "0041acd3-de76-4745-a1ab-c6563f774f33" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.896538] env[63293]: DEBUG oslo_concurrency.lockutils [req-cf5f0b6a-8271-46f5-a196-58dce84f9ec6 req-7ee54622-83eb-474f-a6ca-26a7ebe2d117 service nova] Releasing lock "refresh_cache-97da107a-2e15-4da2-9197-7f85d960ff27" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.915388] env[63293]: DEBUG oslo_vmware.api [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328158, 'name': ReconfigVM_Task, 'duration_secs': 0.38381} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.917766] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Reconfigured VM instance instance-0000005f to attach disk [datastore1] volume-7a0b3a44-ef09-4b4d-9007-6d4753db1906/volume-7a0b3a44-ef09-4b4d-9007-6d4753db1906.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.922854] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3997e7e3-7f2b-4604-b319-8668e73ae565 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.940200] env[63293]: DEBUG oslo_vmware.api [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 1038.940200] env[63293]: value = "task-1328159" [ 1038.940200] env[63293]: _type = "Task" [ 1038.940200] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.953095] env[63293]: DEBUG oslo_vmware.api [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328159, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.011571] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddfc48dd-9f44-4b6a-8dcb-60b4053141e3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.019185] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3341bd82-345b-4095-b0e1-65ed0410f906 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.060788] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d52bd6-6708-4ae6-8fbf-086529c6c4aa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.068427] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca7da3c-8eea-4c94-8d81-26dfbad84df3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.076268] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "b0171e46-ec0b-4fd6-bb04-6f943757b083" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.076524] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "b0171e46-ec0b-4fd6-bb04-6f943757b083" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.085828] env[63293]: DEBUG nova.compute.provider_tree [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.087510] env[63293]: DEBUG nova.compute.manager [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1039.379022] env[63293]: DEBUG nova.compute.manager [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1039.447663] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.448130] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquired lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.448130] env[63293]: DEBUG nova.network.neutron [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1039.454712] env[63293]: DEBUG oslo_vmware.api [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328159, 'name': ReconfigVM_Task, 'duration_secs': 0.147524} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.455232] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283835', 'volume_id': '7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'name': 'volume-7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9937a3c4-c3a6-423b-ac4f-50b7ec5879d2', 'attached_at': '', 'detached_at': '', 'volume_id': '7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'serial': '7a0b3a44-ef09-4b4d-9007-6d4753db1906'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1039.594224] env[63293]: DEBUG nova.scheduler.client.report [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.610227] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.630029] env[63293]: DEBUG oslo_concurrency.lockutils [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.630029] env[63293]: DEBUG oslo_concurrency.lockutils [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.630176] env[63293]: DEBUG oslo_concurrency.lockutils [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.630242] env[63293]: DEBUG oslo_concurrency.lockutils [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.630426] env[63293]: DEBUG oslo_concurrency.lockutils [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.632495] env[63293]: INFO nova.compute.manager [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Terminating instance [ 1039.634236] env[63293]: DEBUG nova.compute.manager [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1039.634431] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1039.635275] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46eda96-d129-496b-93b4-99983cede94e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.643370] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1039.643604] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-984a0945-44ca-4d5a-9d5d-ccb3125914cf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.710502] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1039.710776] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1039.710992] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleting the datastore file [datastore2] 46a9329b-6e3b-454f-8a50-30ff7bc719f0 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1039.711286] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48aac213-7985-496f-9023-3bae361f7ed7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.719706] env[63293]: DEBUG oslo_vmware.api [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for the task: (returnval){ [ 1039.719706] env[63293]: value = "task-1328161" [ 1039.719706] env[63293]: _type = "Task" [ 1039.719706] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.727311] env[63293]: DEBUG oslo_vmware.api [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328161, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.898242] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.100200] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.259s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.101274] env[63293]: DEBUG nova.compute.manager [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1040.103247] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.493s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.104949] env[63293]: INFO nova.compute.claims [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1040.155084] env[63293]: DEBUG nova.network.neutron [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updating instance_info_cache with network_info: [{"id": "4c817793-6a1c-4d84-ac42-0487913a257b", "address": "fa:16:3e:18:84:4b", "network": {"id": "8af42ccf-7ab8-424b-a290-8383c292ef85", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1850566934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eec85c5992d64d62a90e290a1aa2d441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c817793-6a", "ovs_interfaceid": "4c817793-6a1c-4d84-ac42-0487913a257b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.230688] env[63293]: DEBUG oslo_vmware.api [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Task: {'id': task-1328161, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212843} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.231061] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1040.231247] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1040.231439] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1040.231615] env[63293]: INFO nova.compute.manager [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1040.231859] env[63293]: DEBUG oslo.service.loopingcall [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.232128] env[63293]: DEBUG nova.compute.manager [-] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1040.232194] env[63293]: DEBUG nova.network.neutron [-] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1040.493571] env[63293]: DEBUG nova.objects.instance [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'flavor' on Instance uuid 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.545285] env[63293]: DEBUG nova.compute.manager [req-cb9caedd-ecdf-4357-b7a3-200c8100a01d req-644d6ac7-f3ef-4209-b38b-b240fa9348d9 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Received event network-vif-deleted-f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.545285] env[63293]: INFO nova.compute.manager [req-cb9caedd-ecdf-4357-b7a3-200c8100a01d req-644d6ac7-f3ef-4209-b38b-b240fa9348d9 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Neutron deleted interface f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b; detaching it from the instance and deleting it from the info cache [ 1040.545285] env[63293]: DEBUG nova.network.neutron [req-cb9caedd-ecdf-4357-b7a3-200c8100a01d req-644d6ac7-f3ef-4209-b38b-b240fa9348d9 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.609467] env[63293]: DEBUG nova.compute.utils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1040.613330] env[63293]: DEBUG nova.compute.manager [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1040.613595] env[63293]: DEBUG nova.network.neutron [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1040.653177] env[63293]: DEBUG nova.policy [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '53698361c41947a38f1b393e0d5290c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8adfd01c54641fb8bee8cd534d50569', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 1040.657657] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Releasing lock "refresh_cache-f429869f-9e9f-443e-a980-67a80c3bd799" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.658562] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835d6bed-71b4-4adf-9a45-06afda6ca6c9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.666687] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Resuming the VM {{(pid=63293) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1040.666932] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-682a2de4-4923-47e4-83bc-65edcda629ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.674110] env[63293]: DEBUG oslo_vmware.api [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1040.674110] env[63293]: value = "task-1328162" [ 1040.674110] env[63293]: _type = "Task" [ 1040.674110] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.681857] env[63293]: DEBUG oslo_vmware.api [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328162, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.928867] env[63293]: DEBUG nova.network.neutron [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Successfully created port: de6fe4cc-5e56-4066-bee8-f16b1fd6bed4 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1040.999256] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e1fd7b19-a500-403d-96ea-f52913796a3d tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.279s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.022528] env[63293]: DEBUG nova.network.neutron [-] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.047547] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed11343d-a0db-4ac1-ab26-1dfdebfbd850 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.069862] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26490b2b-e326-4145-a932-fd15f1c31db9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.107430] env[63293]: DEBUG nova.compute.manager [req-cb9caedd-ecdf-4357-b7a3-200c8100a01d req-644d6ac7-f3ef-4209-b38b-b240fa9348d9 service nova] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Detach interface failed, port_id=f4c5ccaf-1c23-4eda-a3f5-7e6ec4326a3b, reason: Instance 46a9329b-6e3b-454f-8a50-30ff7bc719f0 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1041.117385] env[63293]: DEBUG nova.compute.manager [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1041.183895] env[63293]: DEBUG oslo_vmware.api [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328162, 'name': PowerOnVM_Task, 'duration_secs': 0.490663} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.185018] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Resumed the VM {{(pid=63293) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1041.185782] env[63293]: DEBUG nova.compute.manager [None req-ea046ca2-748f-42cd-9a2e-634b406334df tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1041.187521] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e03c725-011e-47dd-bbe6-4e17dea7ddce {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.286778] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96507d02-482e-4616-906d-542f591acf91 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.295338] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b1d58b-66ac-4118-b0d0-9b9158d377d6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.299516] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.299838] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.332403] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693e2233-7897-48ba-810d-7bb0477fcd2c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.340939] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3aad726-9550-49de-9521-5813c044e34d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.355742] env[63293]: DEBUG nova.compute.provider_tree [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.524841] env[63293]: INFO nova.compute.manager [-] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Took 1.29 seconds to deallocate network for instance. [ 1041.803911] env[63293]: INFO nova.compute.manager [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Detaching volume 4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6 [ 1041.839247] env[63293]: INFO nova.virt.block_device [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Attempting to driver detach volume 4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6 from mountpoint /dev/sdb [ 1041.839512] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Volume detach. Driver type: vmdk {{(pid=63293) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1041.839710] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283832', 'volume_id': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'name': 'volume-4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9937a3c4-c3a6-423b-ac4f-50b7ec5879d2', 'attached_at': '', 'detached_at': '', 'volume_id': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'serial': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1041.840644] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60dc5003-4f87-4def-91fb-744593797f5b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.866279] env[63293]: DEBUG nova.scheduler.client.report [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.870187] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ac473d-8680-48c6-a792-d22f86e15d0a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.877752] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73093350-6ee0-44ee-80f9-f664bd7accb9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.902681] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0629126-371e-4290-9b8e-3e3627281f64 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.919139] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] The volume has not been displaced from its original location: [datastore1] volume-4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6/volume-4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6.vmdk. No consolidation needed. {{(pid=63293) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1041.924702] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1041.925024] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8494fa84-1f79-4b24-b7ee-f615446d5d39 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.943871] env[63293]: DEBUG oslo_vmware.api [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 1041.943871] env[63293]: value = "task-1328163" [ 1041.943871] env[63293]: _type = "Task" [ 1041.943871] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.952363] env[63293]: DEBUG oslo_vmware.api [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328163, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.031972] env[63293]: DEBUG oslo_concurrency.lockutils [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.126298] env[63293]: DEBUG nova.compute.manager [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1042.153050] env[63293]: DEBUG nova.virt.hardware [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1042.153224] env[63293]: DEBUG nova.virt.hardware [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1042.153282] env[63293]: DEBUG nova.virt.hardware [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1042.153469] env[63293]: DEBUG nova.virt.hardware [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1042.153636] env[63293]: DEBUG nova.virt.hardware [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1042.153790] env[63293]: DEBUG nova.virt.hardware [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1042.154022] env[63293]: DEBUG nova.virt.hardware [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1042.154190] env[63293]: DEBUG nova.virt.hardware [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1042.154378] env[63293]: DEBUG nova.virt.hardware [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1042.154557] env[63293]: DEBUG nova.virt.hardware [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1042.154749] env[63293]: DEBUG nova.virt.hardware [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.155665] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5c66e2-d621-4b48-b34c-37594bb3a285 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.164183] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a123fb54-b251-48bf-866a-132b0ca71b97 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.376021] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.376021] env[63293]: DEBUG nova.compute.manager [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1042.377477] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.479s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.379718] env[63293]: INFO nova.compute.claims [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.454103] env[63293]: DEBUG oslo_vmware.api [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328163, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.469289] env[63293]: DEBUG nova.network.neutron [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Successfully updated port: de6fe4cc-5e56-4066-bee8-f16b1fd6bed4 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1042.568153] env[63293]: DEBUG nova.compute.manager [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Received event network-vif-plugged-de6fe4cc-5e56-4066-bee8-f16b1fd6bed4 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.568389] env[63293]: DEBUG oslo_concurrency.lockutils [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] Acquiring lock "07418025-09fc-4793-a8ea-f059f42251ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.568606] env[63293]: DEBUG oslo_concurrency.lockutils [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] Lock "07418025-09fc-4793-a8ea-f059f42251ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.568804] env[63293]: DEBUG oslo_concurrency.lockutils [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] Lock "07418025-09fc-4793-a8ea-f059f42251ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.568986] env[63293]: DEBUG nova.compute.manager [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] No waiting events found dispatching network-vif-plugged-de6fe4cc-5e56-4066-bee8-f16b1fd6bed4 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1042.569166] env[63293]: WARNING nova.compute.manager [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Received unexpected event network-vif-plugged-de6fe4cc-5e56-4066-bee8-f16b1fd6bed4 for instance with vm_state building and task_state spawning. [ 1042.569394] env[63293]: DEBUG nova.compute.manager [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Received event network-changed-de6fe4cc-5e56-4066-bee8-f16b1fd6bed4 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.569483] env[63293]: DEBUG nova.compute.manager [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Refreshing instance network info cache due to event network-changed-de6fe4cc-5e56-4066-bee8-f16b1fd6bed4. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1042.569668] env[63293]: DEBUG oslo_concurrency.lockutils [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] Acquiring lock "refresh_cache-07418025-09fc-4793-a8ea-f059f42251ea" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.569805] env[63293]: DEBUG oslo_concurrency.lockutils [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] Acquired lock "refresh_cache-07418025-09fc-4793-a8ea-f059f42251ea" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.569967] env[63293]: DEBUG nova.network.neutron [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Refreshing network info cache for port de6fe4cc-5e56-4066-bee8-f16b1fd6bed4 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1042.885110] env[63293]: DEBUG nova.compute.utils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1042.889754] env[63293]: DEBUG nova.compute.manager [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Not allocating networking since 'none' was specified. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1042.954572] env[63293]: DEBUG oslo_vmware.api [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328163, 'name': ReconfigVM_Task, 'duration_secs': 0.647967} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.954869] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1042.959593] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9b20812-dd32-482f-bb5f-cfeb99da1a34 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.971436] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "refresh_cache-07418025-09fc-4793-a8ea-f059f42251ea" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.974801] env[63293]: DEBUG oslo_vmware.api [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 1042.974801] env[63293]: value = "task-1328164" [ 1042.974801] env[63293]: _type = "Task" [ 1042.974801] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.982915] env[63293]: DEBUG oslo_vmware.api [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328164, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.100486] env[63293]: DEBUG nova.network.neutron [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1043.177718] env[63293]: DEBUG nova.network.neutron [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.393696] env[63293]: DEBUG nova.compute.manager [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1043.485045] env[63293]: DEBUG oslo_vmware.api [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328164, 'name': ReconfigVM_Task, 'duration_secs': 0.220592} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.488046] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283832', 'volume_id': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'name': 'volume-4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9937a3c4-c3a6-423b-ac4f-50b7ec5879d2', 'attached_at': '', 'detached_at': '', 'volume_id': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6', 'serial': '4f784ba5-a2a6-46bc-b1e8-688ab7df4ba6'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1043.519470] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7071f2ee-5f17-4326-b5a4-6adc9501ccb5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.526924] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba726d2c-1542-41ff-ab78-7c045482230a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.557349] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed27ffb-9f11-4ca1-ab96-c1e350fa807a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.565177] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53e6b89-39f5-4bd7-a03e-9fe5a3587511 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.578167] env[63293]: DEBUG nova.compute.provider_tree [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.680238] env[63293]: DEBUG oslo_concurrency.lockutils [req-2ce66456-8ab3-4eb5-94a3-b7efdf7329c9 req-4b6005e4-f229-4d3f-ad68-8bc1ea51aee9 service nova] Releasing lock "refresh_cache-07418025-09fc-4793-a8ea-f059f42251ea" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.680616] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired lock "refresh_cache-07418025-09fc-4793-a8ea-f059f42251ea" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.680774] env[63293]: DEBUG nova.network.neutron [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1044.028751] env[63293]: DEBUG nova.objects.instance [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'flavor' on Instance uuid 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.081191] env[63293]: DEBUG nova.scheduler.client.report [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.210545] env[63293]: DEBUG nova.network.neutron [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1044.323295] env[63293]: DEBUG nova.network.neutron [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Updating instance_info_cache with network_info: [{"id": "de6fe4cc-5e56-4066-bee8-f16b1fd6bed4", "address": "fa:16:3e:5c:74:a3", "network": {"id": "69a5064b-e3a0-4ef6-b722-09fbb9b50ca8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1786908568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b8adfd01c54641fb8bee8cd534d50569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6fe4cc-5e", "ovs_interfaceid": "de6fe4cc-5e56-4066-bee8-f16b1fd6bed4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.403643] env[63293]: DEBUG nova.compute.manager [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1044.428338] env[63293]: DEBUG nova.virt.hardware [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.428583] env[63293]: DEBUG nova.virt.hardware [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.428743] env[63293]: DEBUG nova.virt.hardware [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.428927] env[63293]: DEBUG nova.virt.hardware [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.429092] env[63293]: DEBUG nova.virt.hardware [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.429248] env[63293]: DEBUG nova.virt.hardware [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.429519] env[63293]: DEBUG nova.virt.hardware [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.429693] env[63293]: DEBUG nova.virt.hardware [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.429864] env[63293]: DEBUG nova.virt.hardware [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.430041] env[63293]: DEBUG nova.virt.hardware [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.430219] env[63293]: DEBUG nova.virt.hardware [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.431113] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bb7a40-bf99-47e9-b1f4-cd50f2e994a1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.439206] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1140ff-b354-4ff3-91b1-16d0200be724 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.453577] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.459819] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Creating folder: Project (04a7f8d5468d4e85b49904dd5c01aaec). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1044.460203] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ed06426-cd35-493d-8d16-8eb352ea3692 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.470754] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Created folder: Project (04a7f8d5468d4e85b49904dd5c01aaec) in parent group-v283678. [ 1044.471061] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Creating folder: Instances. Parent ref: group-v283836. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1044.471379] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69b5a489-92af-4f5c-b767-684218e1462b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.480987] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Created folder: Instances in parent group-v283836. [ 1044.481281] env[63293]: DEBUG oslo.service.loopingcall [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.481478] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1044.481686] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7322d79d-46ad-4026-8948-ab3f7f46d631 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.497520] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.497520] env[63293]: value = "task-1328167" [ 1044.497520] env[63293]: _type = "Task" [ 1044.497520] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.504698] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328167, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.585696] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.208s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.586244] env[63293]: DEBUG nova.compute.manager [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1044.588960] env[63293]: DEBUG oslo_concurrency.lockutils [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.557s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.589192] env[63293]: DEBUG nova.objects.instance [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lazy-loading 'resources' on Instance uuid 46a9329b-6e3b-454f-8a50-30ff7bc719f0 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.826025] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Releasing lock "refresh_cache-07418025-09fc-4793-a8ea-f059f42251ea" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.826904] env[63293]: DEBUG nova.compute.manager [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Instance network_info: |[{"id": "de6fe4cc-5e56-4066-bee8-f16b1fd6bed4", "address": "fa:16:3e:5c:74:a3", "network": {"id": "69a5064b-e3a0-4ef6-b722-09fbb9b50ca8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1786908568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b8adfd01c54641fb8bee8cd534d50569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6fe4cc-5e", "ovs_interfaceid": "de6fe4cc-5e56-4066-bee8-f16b1fd6bed4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1044.826904] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:74:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bff6c3a1-cc80-46ca-86c0-6dbb029edddb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'de6fe4cc-5e56-4066-bee8-f16b1fd6bed4', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.834660] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Creating folder: Project (b8adfd01c54641fb8bee8cd534d50569). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1044.834954] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f548c585-d563-4fc9-a439-f578706c3355 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.845495] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Created folder: Project (b8adfd01c54641fb8bee8cd534d50569) in parent group-v283678. [ 1044.845687] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Creating folder: Instances. Parent ref: group-v283839. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1044.845928] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f76a2f72-a0f6-4a61-a7ee-bea2a4363363 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.855923] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Created folder: Instances in parent group-v283839. [ 1044.856167] env[63293]: DEBUG oslo.service.loopingcall [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.856359] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1044.856558] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1802c523-3a0c-4f7f-8240-327cd59c0ebd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.876132] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.876132] env[63293]: value = "task-1328170" [ 1044.876132] env[63293]: _type = "Task" [ 1044.876132] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.884222] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328170, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.007610] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328167, 'name': CreateVM_Task, 'duration_secs': 0.293709} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.007770] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1045.008233] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.008405] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.008731] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1045.009018] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dbe86a0-4627-48fb-bbb2-6285ce9e6bba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.014602] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1045.014602] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5237c97b-f78b-338a-daa4-856e1bd8cfdc" [ 1045.014602] env[63293]: _type = "Task" [ 1045.014602] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.023602] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5237c97b-f78b-338a-daa4-856e1bd8cfdc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.036194] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5f327f15-359c-4e77-b574-1cbaf13b2237 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.736s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.092347] env[63293]: DEBUG nova.compute.utils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1045.097433] env[63293]: DEBUG nova.compute.manager [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Not allocating networking since 'none' was specified. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1045.227394] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbaecd48-aa76-4444-b19d-b4ff4e2a22b1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.236278] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d7b900-1b2f-445a-a194-2f092b309813 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.265744] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562594ed-86ff-4d7e-a910-718434991dab {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.272607] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d551767d-0b2f-4c91-827c-cf394965377a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.285157] env[63293]: DEBUG nova.compute.provider_tree [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.386655] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328170, 'name': CreateVM_Task, 'duration_secs': 0.315006} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.386816] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1045.387433] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.525351] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5237c97b-f78b-338a-daa4-856e1bd8cfdc, 'name': SearchDatastore_Task, 'duration_secs': 0.008636} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.525651] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.525889] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1045.526138] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.526293] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.526472] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1045.526739] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.527054] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1045.527283] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6af42272-f19d-4e87-9262-e96855709072 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.528892] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86f096d1-43ce-4f60-b219-eae5a2d5b879 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.534051] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1045.534051] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fc35fc-fcbc-0006-3094-a4fabcd545e3" [ 1045.534051] env[63293]: _type = "Task" [ 1045.534051] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.537779] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1045.537966] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1045.539103] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-331e850a-dfb0-4dbb-93e7-d49a2131067a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.543677] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fc35fc-fcbc-0006-3094-a4fabcd545e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.546548] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1045.546548] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52402836-ffce-54ea-1af3-0c6cc32a1719" [ 1045.546548] env[63293]: _type = "Task" [ 1045.546548] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.553592] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52402836-ffce-54ea-1af3-0c6cc32a1719, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.588159] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.588448] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.597289] env[63293]: DEBUG nova.compute.manager [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1045.622197] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "f429869f-9e9f-443e-a980-67a80c3bd799" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.622474] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "f429869f-9e9f-443e-a980-67a80c3bd799" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.622707] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.622902] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.623086] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "f429869f-9e9f-443e-a980-67a80c3bd799-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.625146] env[63293]: INFO nova.compute.manager [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Terminating instance [ 1045.627182] env[63293]: DEBUG nova.compute.manager [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1045.627383] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1045.628293] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3df894e-5cf1-43b4-a272-4f0403c44c39 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.636188] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1045.636412] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49bd68be-545a-40b0-813a-95731d8a0e90 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.642845] env[63293]: DEBUG oslo_vmware.api [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1045.642845] env[63293]: value = "task-1328171" [ 1045.642845] env[63293]: _type = "Task" [ 1045.642845] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.651152] env[63293]: DEBUG oslo_vmware.api [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328171, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.788049] env[63293]: DEBUG nova.scheduler.client.report [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.044682] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52fc35fc-fcbc-0006-3094-a4fabcd545e3, 'name': SearchDatastore_Task, 'duration_secs': 0.010553} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.045011] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.045259] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1046.045471] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.054628] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52402836-ffce-54ea-1af3-0c6cc32a1719, 'name': SearchDatastore_Task, 'duration_secs': 0.007594} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.055367] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86a808bc-852e-4b67-89b2-5d71cde4daf4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.060375] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1046.060375] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f8223b-4a9e-ff2a-5983-5c3c242060d3" [ 1046.060375] env[63293]: _type = "Task" [ 1046.060375] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.068037] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f8223b-4a9e-ff2a-5983-5c3c242060d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.091878] env[63293]: INFO nova.compute.manager [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Detaching volume 7a0b3a44-ef09-4b4d-9007-6d4753db1906 [ 1046.127158] env[63293]: INFO nova.virt.block_device [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Attempting to driver detach volume 7a0b3a44-ef09-4b4d-9007-6d4753db1906 from mountpoint /dev/sdc [ 1046.127412] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Volume detach. Driver type: vmdk {{(pid=63293) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1046.127605] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283835', 'volume_id': '7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'name': 'volume-7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9937a3c4-c3a6-423b-ac4f-50b7ec5879d2', 'attached_at': '', 'detached_at': '', 'volume_id': '7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'serial': '7a0b3a44-ef09-4b4d-9007-6d4753db1906'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1046.128552] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73f9350-a5d0-4ea9-9c2f-c4bb24c7e12d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.153081] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3c6176-92e1-411b-94c5-33fca76b5948 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.162592] env[63293]: DEBUG oslo_vmware.api [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328171, 'name': PowerOffVM_Task, 'duration_secs': 0.194479} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.164171] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1046.164360] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1046.164929] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8be5586-7882-4797-b814-f987766c17ae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.166637] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04ea105-a885-4aa2-b801-02bf07623371 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.189076] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc06903c-9dfc-4830-8162-47f7b95538e8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.203544] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] The volume has not been displaced from its original location: [datastore1] volume-7a0b3a44-ef09-4b4d-9007-6d4753db1906/volume-7a0b3a44-ef09-4b4d-9007-6d4753db1906.vmdk. No consolidation needed. {{(pid=63293) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1046.208703] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Reconfiguring VM instance instance-0000005f to detach disk 2002 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1046.209235] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70980899-796d-4f42-a52d-bc087220a734 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.226922] env[63293]: DEBUG oslo_vmware.api [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 1046.226922] env[63293]: value = "task-1328173" [ 1046.226922] env[63293]: _type = "Task" [ 1046.226922] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.234589] env[63293]: DEBUG oslo_vmware.api [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328173, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.251984] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1046.252224] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1046.252414] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Deleting the datastore file [datastore1] f429869f-9e9f-443e-a980-67a80c3bd799 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.252674] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d69fa167-3806-4f39-be30-c332ab7ae7be {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.259898] env[63293]: DEBUG oslo_vmware.api [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for the task: (returnval){ [ 1046.259898] env[63293]: value = "task-1328174" [ 1046.259898] env[63293]: _type = "Task" [ 1046.259898] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.267839] env[63293]: DEBUG oslo_vmware.api [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328174, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.293824] env[63293]: DEBUG oslo_concurrency.lockutils [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.705s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.311675] env[63293]: INFO nova.scheduler.client.report [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Deleted allocations for instance 46a9329b-6e3b-454f-8a50-30ff7bc719f0 [ 1046.571140] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f8223b-4a9e-ff2a-5983-5c3c242060d3, 'name': SearchDatastore_Task, 'duration_secs': 0.024633} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.571434] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.571695] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] b0171e46-ec0b-4fd6-bb04-6f943757b083/b0171e46-ec0b-4fd6-bb04-6f943757b083.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1046.571984] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.572412] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.572451] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c4829653-0f41-4d56-bb1b-f9edaf9931a6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.574509] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ff27d60-864f-4120-a36e-56c5039f0ee0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.580741] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1046.580741] env[63293]: value = "task-1328175" [ 1046.580741] env[63293]: _type = "Task" [ 1046.580741] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.584417] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.584593] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1046.585580] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf9a85a6-23c7-4097-a118-d2335d697ef9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.590397] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328175, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.593181] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1046.593181] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529c8a6d-1be3-9fa2-e490-29705dd606c0" [ 1046.593181] env[63293]: _type = "Task" [ 1046.593181] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.600169] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529c8a6d-1be3-9fa2-e490-29705dd606c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.606307] env[63293]: DEBUG nova.compute.manager [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1046.632720] env[63293]: DEBUG nova.virt.hardware [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1046.632966] env[63293]: DEBUG nova.virt.hardware [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1046.633143] env[63293]: DEBUG nova.virt.hardware [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1046.633334] env[63293]: DEBUG nova.virt.hardware [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1046.633516] env[63293]: DEBUG nova.virt.hardware [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1046.633629] env[63293]: DEBUG nova.virt.hardware [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1046.633844] env[63293]: DEBUG nova.virt.hardware [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1046.634017] env[63293]: DEBUG nova.virt.hardware [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1046.634199] env[63293]: DEBUG nova.virt.hardware [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1046.634365] env[63293]: DEBUG nova.virt.hardware [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1046.634542] env[63293]: DEBUG nova.virt.hardware [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1046.635455] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-580dec90-a19f-4578-b366-efbdd742d3f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.643105] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c9ef6d3-c4cc-4b9d-8e57-bfcfa896cd2a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.657686] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1046.663158] env[63293]: DEBUG oslo.service.loopingcall [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.663397] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1046.663607] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7badff3-9375-4aef-a716-1910f109a7f0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.679596] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1046.679596] env[63293]: value = "task-1328176" [ 1046.679596] env[63293]: _type = "Task" [ 1046.679596] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.686882] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328176, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.739429] env[63293]: DEBUG oslo_vmware.api [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328173, 'name': ReconfigVM_Task, 'duration_secs': 0.219157} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.739718] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Reconfigured VM instance instance-0000005f to detach disk 2002 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1046.744843] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebc78fcd-10e8-420b-81af-bab6fe56d901 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.765646] env[63293]: DEBUG oslo_vmware.api [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 1046.765646] env[63293]: value = "task-1328177" [ 1046.765646] env[63293]: _type = "Task" [ 1046.765646] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.772583] env[63293]: DEBUG oslo_vmware.api [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Task: {'id': task-1328174, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188936} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.773308] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1046.773522] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1046.773786] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1046.773896] env[63293]: INFO nova.compute.manager [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1046.774195] env[63293]: DEBUG oslo.service.loopingcall [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.774422] env[63293]: DEBUG nova.compute.manager [-] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1046.774518] env[63293]: DEBUG nova.network.neutron [-] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1046.780159] env[63293]: DEBUG oslo_vmware.api [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328177, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.821677] env[63293]: DEBUG oslo_concurrency.lockutils [None req-af184fdb-6d00-45e4-9a2a-e64bc1066323 tempest-ServersTestJSON-1254883005 tempest-ServersTestJSON-1254883005-project-member] Lock "46a9329b-6e3b-454f-8a50-30ff7bc719f0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.191s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.088471] env[63293]: DEBUG nova.compute.manager [req-d007e2ad-5fcb-4232-9579-630304eb59b0 req-4f154145-8b87-46f5-ab56-d299d23fb6f3 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Received event network-vif-deleted-4c817793-6a1c-4d84-ac42-0487913a257b {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.088726] env[63293]: INFO nova.compute.manager [req-d007e2ad-5fcb-4232-9579-630304eb59b0 req-4f154145-8b87-46f5-ab56-d299d23fb6f3 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Neutron deleted interface 4c817793-6a1c-4d84-ac42-0487913a257b; detaching it from the instance and deleting it from the info cache [ 1047.089028] env[63293]: DEBUG nova.network.neutron [req-d007e2ad-5fcb-4232-9579-630304eb59b0 req-4f154145-8b87-46f5-ab56-d299d23fb6f3 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.094073] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328175, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458671} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.097410] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] b0171e46-ec0b-4fd6-bb04-6f943757b083/b0171e46-ec0b-4fd6-bb04-6f943757b083.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1047.097787] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1047.098109] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0f251dd-b3de-4675-a2d4-24fd8f6390ff {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.106734] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529c8a6d-1be3-9fa2-e490-29705dd606c0, 'name': SearchDatastore_Task, 'duration_secs': 0.007899} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.108696] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1047.108696] env[63293]: value = "task-1328178" [ 1047.108696] env[63293]: _type = "Task" [ 1047.108696] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.108929] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d70d4d76-1fde-42d8-a573-4043d5eaf0cf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.117538] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1047.117538] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5217bd46-ebd7-3022-f666-f32945e233c1" [ 1047.117538] env[63293]: _type = "Task" [ 1047.117538] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.120829] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328178, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.128707] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5217bd46-ebd7-3022-f666-f32945e233c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.189590] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328176, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.277066] env[63293]: DEBUG oslo_vmware.api [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328177, 'name': ReconfigVM_Task, 'duration_secs': 0.15541} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.277066] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283835', 'volume_id': '7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'name': 'volume-7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9937a3c4-c3a6-423b-ac4f-50b7ec5879d2', 'attached_at': '', 'detached_at': '', 'volume_id': '7a0b3a44-ef09-4b4d-9007-6d4753db1906', 'serial': '7a0b3a44-ef09-4b4d-9007-6d4753db1906'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1047.567709] env[63293]: DEBUG nova.network.neutron [-] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.595558] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6dcaa9a3-823e-4ef8-9ca6-90b0d3555dc1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.606822] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f60949-e2f9-401c-a882-583af3bb7b86 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.628194] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328178, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06515} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.631632] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1047.631941] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5217bd46-ebd7-3022-f666-f32945e233c1, 'name': SearchDatastore_Task, 'duration_secs': 0.010943} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.640056] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f454a053-dbf7-4ede-88c7-c31b3fc380e2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.642513] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.642733] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 07418025-09fc-4793-a8ea-f059f42251ea/07418025-09fc-4793-a8ea-f059f42251ea.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1047.643924] env[63293]: DEBUG nova.compute.manager [req-d007e2ad-5fcb-4232-9579-630304eb59b0 req-4f154145-8b87-46f5-ab56-d299d23fb6f3 service nova] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Detach interface failed, port_id=4c817793-6a1c-4d84-ac42-0487913a257b, reason: Instance f429869f-9e9f-443e-a980-67a80c3bd799 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1047.643924] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43d8ec5e-d9e4-4d4c-8e2a-833572e4910a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.662729] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] b0171e46-ec0b-4fd6-bb04-6f943757b083/b0171e46-ec0b-4fd6-bb04-6f943757b083.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.663684] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ea198e7-0f3f-4922-9a32-db12367ecdc4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.677588] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1047.677588] env[63293]: value = "task-1328179" [ 1047.677588] env[63293]: _type = "Task" [ 1047.677588] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.685096] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1047.685096] env[63293]: value = "task-1328180" [ 1047.685096] env[63293]: _type = "Task" [ 1047.685096] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.694732] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328179, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.694932] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328176, 'name': CreateVM_Task, 'duration_secs': 0.557576} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.695416] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1047.695850] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.696096] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.696532] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.696876] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1480cc52-474f-428f-ac5e-746a5f45413a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.701316] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328180, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.704199] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1047.704199] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52238883-6983-a0dd-b798-95b1c77fd525" [ 1047.704199] env[63293]: _type = "Task" [ 1047.704199] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.712293] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52238883-6983-a0dd-b798-95b1c77fd525, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.821726] env[63293]: DEBUG nova.objects.instance [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'flavor' on Instance uuid 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.070892] env[63293]: INFO nova.compute.manager [-] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Took 1.30 seconds to deallocate network for instance. [ 1048.189618] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328179, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53597} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.192935] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 07418025-09fc-4793-a8ea-f059f42251ea/07418025-09fc-4793-a8ea-f059f42251ea.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1048.193207] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1048.193478] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-233cfc85-4c67-4c11-9227-6d817cebca10 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.202142] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328180, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.203457] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1048.203457] env[63293]: value = "task-1328181" [ 1048.203457] env[63293]: _type = "Task" [ 1048.203457] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.214666] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328181, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.218047] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52238883-6983-a0dd-b798-95b1c77fd525, 'name': SearchDatastore_Task, 'duration_secs': 0.012169} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.218362] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.218599] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1048.218834] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.218983] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.219178] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1048.219446] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd052cbe-48a9-4cde-9361-1d7c3b998847 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.228050] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1048.228240] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1048.228995] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e05e93b-a1b4-442c-a83c-84cbbbc2d1dc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.235432] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1048.235432] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e2dd81-8d37-125e-3412-f3d821a66e67" [ 1048.235432] env[63293]: _type = "Task" [ 1048.235432] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.244645] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e2dd81-8d37-125e-3412-f3d821a66e67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.578489] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.578909] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.579289] env[63293]: DEBUG nova.objects.instance [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lazy-loading 'resources' on Instance uuid f429869f-9e9f-443e-a980-67a80c3bd799 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.701880] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328180, 'name': ReconfigVM_Task, 'duration_secs': 0.579861} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.702255] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Reconfigured VM instance instance-0000006a to attach disk [datastore2] b0171e46-ec0b-4fd6-bb04-6f943757b083/b0171e46-ec0b-4fd6-bb04-6f943757b083.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1048.702938] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10a4c84e-afb1-4253-9dae-49f964c2f21e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.715068] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1048.715068] env[63293]: value = "task-1328182" [ 1048.715068] env[63293]: _type = "Task" [ 1048.715068] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.715287] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328181, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069169} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.715859] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1048.719268] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b3c06b-61b5-48e0-af9c-4bf71a18f564 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.727484] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328182, 'name': Rename_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.744783] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 07418025-09fc-4793-a8ea-f059f42251ea/07418025-09fc-4793-a8ea-f059f42251ea.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.747873] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87dc578b-b681-404e-8075-5599431a45b1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.767399] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e2dd81-8d37-125e-3412-f3d821a66e67, 'name': SearchDatastore_Task, 'duration_secs': 0.009975} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.769104] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1048.769104] env[63293]: value = "task-1328183" [ 1048.769104] env[63293]: _type = "Task" [ 1048.769104] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.769313] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5937a210-2f22-4e3d-a071-1ab38ff6dda5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.777264] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1048.777264] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ddeda4-6259-f033-02d5-1ea2bf143e29" [ 1048.777264] env[63293]: _type = "Task" [ 1048.777264] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.780911] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328183, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.788660] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ddeda4-6259-f033-02d5-1ea2bf143e29, 'name': SearchDatastore_Task, 'duration_secs': 0.009226} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.788930] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.789203] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 0041acd3-de76-4745-a1ab-c6563f774f33/0041acd3-de76-4745-a1ab-c6563f774f33.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1048.789450] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fae402e8-bafc-44f8-881d-578e5cb4b72e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.795276] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1048.795276] env[63293]: value = "task-1328184" [ 1048.795276] env[63293]: _type = "Task" [ 1048.795276] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.802437] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328184, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.828339] env[63293]: DEBUG oslo_concurrency.lockutils [None req-55682ce5-7116-4ea6-ad0a-15289d9066b4 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.240s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.229972] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328182, 'name': Rename_Task, 'duration_secs': 0.127176} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.230438] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1049.230563] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90bf71b8-8d63-4828-bbb1-0eb0dfa26c20 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.235774] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c67ec4-15a0-4a0d-8736-e7ba97ac8c79 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.239971] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1049.239971] env[63293]: value = "task-1328185" [ 1049.239971] env[63293]: _type = "Task" [ 1049.239971] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.246668] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e785a33-ded7-4f42-a21f-92c0fdb38efa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.253339] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328185, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.286126] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ccff97-7994-4bd8-96bf-c20b76040b8e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.297460] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328183, 'name': ReconfigVM_Task, 'duration_secs': 0.391632} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.303264] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 07418025-09fc-4793-a8ea-f059f42251ea/07418025-09fc-4793-a8ea-f059f42251ea.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.305365] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-32955279-87ff-422e-bfa5-639f3c467fbe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.307773] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f42541-54aa-46d2-9fae-b9ba4b62dfcc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.318427] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328184, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.328637] env[63293]: DEBUG nova.compute.provider_tree [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.330280] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1049.330280] env[63293]: value = "task-1328186" [ 1049.330280] env[63293]: _type = "Task" [ 1049.330280] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.340737] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328186, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.750776] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328185, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.793349] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquiring lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.793581] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.793796] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquiring lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.793979] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.794169] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.796321] env[63293]: INFO nova.compute.manager [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Terminating instance [ 1049.800933] env[63293]: DEBUG nova.compute.manager [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1049.801164] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1049.801910] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07993c6-a81d-44be-b943-27bc5f98fca8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.809839] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328184, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543641} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.811770] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 0041acd3-de76-4745-a1ab-c6563f774f33/0041acd3-de76-4745-a1ab-c6563f774f33.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1049.811988] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.812267] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1049.812466] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-483a5318-e512-4179-afed-41d8ab600e92 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.814082] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-685679b7-6290-4d67-8953-2084fbdbe90f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.820365] env[63293]: DEBUG oslo_vmware.api [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for the task: (returnval){ [ 1049.820365] env[63293]: value = "task-1328188" [ 1049.820365] env[63293]: _type = "Task" [ 1049.820365] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.821518] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1049.821518] env[63293]: value = "task-1328187" [ 1049.821518] env[63293]: _type = "Task" [ 1049.821518] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.831402] env[63293]: DEBUG oslo_vmware.api [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328188, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.837079] env[63293]: DEBUG nova.scheduler.client.report [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.840077] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328187, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.845476] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328186, 'name': Rename_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.020265] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.020548] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.020766] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.021037] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.021297] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.023580] env[63293]: INFO nova.compute.manager [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Terminating instance [ 1050.025351] env[63293]: DEBUG nova.compute.manager [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1050.025549] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1050.026462] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afda1138-0d59-484f-8861-86da7c183db5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.035032] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1050.035274] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a242c05-b221-4773-b4a2-feba3061ccfb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.043566] env[63293]: DEBUG oslo_vmware.api [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 1050.043566] env[63293]: value = "task-1328189" [ 1050.043566] env[63293]: _type = "Task" [ 1050.043566] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.055706] env[63293]: DEBUG oslo_vmware.api [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.257722] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328185, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.337821] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328187, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.244391} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.341145] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1050.341888] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.763s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.343702] env[63293]: DEBUG oslo_vmware.api [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328188, 'name': PowerOffVM_Task, 'duration_secs': 0.415745} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.344952] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44826ac-8c06-41a2-b8f3-aead19a26f2d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.347264] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1050.347443] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1050.350848] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3975fc29-1cf2-4417-9df2-693c69bf6fe0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.352464] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328186, 'name': Rename_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.374062] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 0041acd3-de76-4745-a1ab-c6563f774f33/0041acd3-de76-4745-a1ab-c6563f774f33.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.375105] env[63293]: INFO nova.scheduler.client.report [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Deleted allocations for instance f429869f-9e9f-443e-a980-67a80c3bd799 [ 1050.378043] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e367a24b-b619-413d-bd62-04541fea5a0d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.399820] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1050.399820] env[63293]: value = "task-1328191" [ 1050.399820] env[63293]: _type = "Task" [ 1050.399820] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.408340] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328191, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.439799] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1050.440287] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1050.440408] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Deleting the datastore file [datastore2] 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.440962] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e417ea43-9efa-4521-aace-5a7c9598e9d6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.450016] env[63293]: DEBUG oslo_vmware.api [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for the task: (returnval){ [ 1050.450016] env[63293]: value = "task-1328192" [ 1050.450016] env[63293]: _type = "Task" [ 1050.450016] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.459129] env[63293]: DEBUG oslo_vmware.api [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328192, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.558145] env[63293]: DEBUG oslo_vmware.api [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328189, 'name': PowerOffVM_Task, 'duration_secs': 0.240115} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.558670] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1050.558861] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1050.559205] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf09dd86-80f2-4f35-a139-99975b4bd321 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.633806] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1050.633806] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1050.634033] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Deleting the datastore file [datastore1] 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.634344] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d9f38aa6-abd4-45eb-8bc9-a5d66d4cec66 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.643040] env[63293]: DEBUG oslo_vmware.api [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for the task: (returnval){ [ 1050.643040] env[63293]: value = "task-1328194" [ 1050.643040] env[63293]: _type = "Task" [ 1050.643040] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.651777] env[63293]: DEBUG oslo_vmware.api [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.753639] env[63293]: DEBUG oslo_vmware.api [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328185, 'name': PowerOnVM_Task, 'duration_secs': 1.052354} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.754194] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1050.754601] env[63293]: INFO nova.compute.manager [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Took 6.35 seconds to spawn the instance on the hypervisor. [ 1050.754946] env[63293]: DEBUG nova.compute.manager [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.756992] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e13845-fa9b-4865-b475-c5e2a94aadc1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.842893] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328186, 'name': Rename_Task, 'duration_secs': 1.158313} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.842893] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.842893] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-818a0c7e-eafd-4eb0-9d4b-7f1ff1c25f10 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.852028] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1050.852028] env[63293]: value = "task-1328195" [ 1050.852028] env[63293]: _type = "Task" [ 1050.852028] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.860274] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328195, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.898663] env[63293]: DEBUG oslo_concurrency.lockutils [None req-7b2a68b0-9044-499b-86d7-2b360350fbf7 tempest-ServersNegativeTestJSON-1619605080 tempest-ServersNegativeTestJSON-1619605080-project-member] Lock "f429869f-9e9f-443e-a980-67a80c3bd799" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.276s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.911053] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.962228] env[63293]: DEBUG oslo_vmware.api [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Task: {'id': task-1328192, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.428542} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.962699] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.963059] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1050.963791] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1050.963791] env[63293]: INFO nova.compute.manager [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1050.964163] env[63293]: DEBUG oslo.service.loopingcall [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.964468] env[63293]: DEBUG nova.compute.manager [-] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1050.964645] env[63293]: DEBUG nova.network.neutron [-] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1051.155460] env[63293]: DEBUG oslo_vmware.api [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Task: {'id': task-1328194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.442648} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.155765] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1051.155986] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1051.156199] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1051.156407] env[63293]: INFO nova.compute.manager [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1051.156679] env[63293]: DEBUG oslo.service.loopingcall [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1051.156976] env[63293]: DEBUG nova.compute.manager [-] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1051.157059] env[63293]: DEBUG nova.network.neutron [-] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1051.277680] env[63293]: INFO nova.compute.manager [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Took 11.69 seconds to build instance. [ 1051.366227] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328195, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.415981] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328191, 'name': ReconfigVM_Task, 'duration_secs': 0.771747} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.416302] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 0041acd3-de76-4745-a1ab-c6563f774f33/0041acd3-de76-4745-a1ab-c6563f774f33.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.416939] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b808299-88be-4a07-b01a-cf708999379a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.424852] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1051.424852] env[63293]: value = "task-1328196" [ 1051.424852] env[63293]: _type = "Task" [ 1051.424852] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.435656] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328196, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.782676] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a1d4f53a-7e15-4090-8f18-2bdef60263e8 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "b0171e46-ec0b-4fd6-bb04-6f943757b083" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.706s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.864787] env[63293]: DEBUG oslo_vmware.api [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328195, 'name': PowerOnVM_Task, 'duration_secs': 0.545035} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.865352] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1051.865902] env[63293]: INFO nova.compute.manager [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Took 9.74 seconds to spawn the instance on the hypervisor. [ 1051.866562] env[63293]: DEBUG nova.compute.manager [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.867969] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c93dbe4-527b-47e5-a84f-5a04bedc6dd8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.936469] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328196, 'name': Rename_Task, 'duration_secs': 0.337786} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.936734] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1051.937743] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee16e66b-26bc-4462-9b52-01553579390a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.946812] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1051.946812] env[63293]: value = "task-1328197" [ 1051.946812] env[63293]: _type = "Task" [ 1051.946812] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.961803] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328197, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.074541] env[63293]: INFO nova.compute.manager [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Rebuilding instance [ 1052.143023] env[63293]: DEBUG nova.compute.manager [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.143023] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cba31bf-a9d1-4fd7-9b20-f14a23923454 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.206455] env[63293]: DEBUG nova.compute.manager [req-afd761d7-4ad8-4aa5-a2c0-912b57561d5d req-92a68419-1a42-4af8-abbe-efd29509ec2c service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Received event network-vif-deleted-8147f046-74b1-4a9e-b437-02f3a5ee8fe3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1052.206660] env[63293]: INFO nova.compute.manager [req-afd761d7-4ad8-4aa5-a2c0-912b57561d5d req-92a68419-1a42-4af8-abbe-efd29509ec2c service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Neutron deleted interface 8147f046-74b1-4a9e-b437-02f3a5ee8fe3; detaching it from the instance and deleting it from the info cache [ 1052.206833] env[63293]: DEBUG nova.network.neutron [req-afd761d7-4ad8-4aa5-a2c0-912b57561d5d req-92a68419-1a42-4af8-abbe-efd29509ec2c service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.402861] env[63293]: INFO nova.compute.manager [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Took 14.58 seconds to build instance. [ 1052.459399] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328197, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.658572] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1052.659104] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70d8cbef-262e-483b-98ce-4bb8f944570c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.669629] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1052.669629] env[63293]: value = "task-1328198" [ 1052.669629] env[63293]: _type = "Task" [ 1052.669629] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.680335] env[63293]: DEBUG nova.network.neutron [-] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.681717] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328198, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.711736] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c49c0927-d80e-4d30-ba72-72a304c9296b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.733293] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e76afd-f993-4385-80c6-a0a8e2463db6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.775908] env[63293]: DEBUG nova.compute.manager [req-afd761d7-4ad8-4aa5-a2c0-912b57561d5d req-92a68419-1a42-4af8-abbe-efd29509ec2c service nova] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Detach interface failed, port_id=8147f046-74b1-4a9e-b437-02f3a5ee8fe3, reason: Instance 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1052.828680] env[63293]: DEBUG nova.network.neutron [-] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.906457] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f25b7668-9ae4-4bb3-8ed5-19b059a7b18d tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "07418025-09fc-4793-a8ea-f059f42251ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.096s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.959597] env[63293]: DEBUG oslo_vmware.api [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328197, 'name': PowerOnVM_Task, 'duration_secs': 0.987003} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.959886] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1052.960100] env[63293]: INFO nova.compute.manager [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Took 6.35 seconds to spawn the instance on the hypervisor. [ 1052.960318] env[63293]: DEBUG nova.compute.manager [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.961176] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f221cb2-f3f5-4ef2-8037-5973177da7e7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.180348] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328198, 'name': PowerOffVM_Task, 'duration_secs': 0.118074} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.180591] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1053.180811] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1053.181663] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1adea70-9ccf-4ab8-bd35-e63a53922985 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.184358] env[63293]: INFO nova.compute.manager [-] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Took 2.22 seconds to deallocate network for instance. [ 1053.191070] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1053.191339] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c27bb76-88d2-45e8-a09a-cd5c2acc969c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.222133] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1053.222133] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1053.222342] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Deleting the datastore file [datastore2] b0171e46-ec0b-4fd6-bb04-6f943757b083 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.222899] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a47cb46-3429-4637-a1ed-fe15e3269448 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.230575] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1053.230575] env[63293]: value = "task-1328200" [ 1053.230575] env[63293]: _type = "Task" [ 1053.230575] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.239602] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328200, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.331230] env[63293]: INFO nova.compute.manager [-] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Took 2.17 seconds to deallocate network for instance. [ 1053.442172] env[63293]: INFO nova.compute.manager [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Rescuing [ 1053.442476] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "refresh_cache-07418025-09fc-4793-a8ea-f059f42251ea" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.442578] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired lock "refresh_cache-07418025-09fc-4793-a8ea-f059f42251ea" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.442746] env[63293]: DEBUG nova.network.neutron [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1053.478738] env[63293]: INFO nova.compute.manager [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Took 13.60 seconds to build instance. [ 1053.693337] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.693637] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.693868] env[63293]: DEBUG nova.objects.instance [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Lazy-loading 'resources' on Instance uuid 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.741804] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328200, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.837126] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.981564] env[63293]: DEBUG oslo_concurrency.lockutils [None req-56ddcb48-a2bb-4397-bbc7-925b5dfdab26 tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "0041acd3-de76-4745-a1ab-c6563f774f33" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.106s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.141462] env[63293]: DEBUG nova.network.neutron [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Updating instance_info_cache with network_info: [{"id": "de6fe4cc-5e56-4066-bee8-f16b1fd6bed4", "address": "fa:16:3e:5c:74:a3", "network": {"id": "69a5064b-e3a0-4ef6-b722-09fbb9b50ca8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1786908568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b8adfd01c54641fb8bee8cd534d50569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde6fe4cc-5e", "ovs_interfaceid": "de6fe4cc-5e56-4066-bee8-f16b1fd6bed4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.239072] env[63293]: DEBUG nova.compute.manager [req-bd6f62da-f0c7-4f27-9171-c045e2dacfc9 req-6cf7b329-d683-4335-b539-e544c09ef366 service nova] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Received event network-vif-deleted-cc516442-7d3c-4f2e-be0a-8c9e2b0bf6d3 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1054.245325] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328200, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.537203} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.247812] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1054.248015] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1054.248209] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1054.317286] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0cda5f4-a8b2-4bbb-acbb-8e2f92875d06 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.325383] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7452bc7-6cad-4221-803f-13214dd60075 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.356785] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4063ac35-e3f1-47fd-8e86-35e0cce5c2db {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.364506] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedb7b82-0caa-48f0-934e-08a560ec612d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.378611] env[63293]: DEBUG nova.compute.provider_tree [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.644265] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Releasing lock "refresh_cache-07418025-09fc-4793-a8ea-f059f42251ea" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.882173] env[63293]: DEBUG nova.scheduler.client.report [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.176410] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1055.176720] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85f50736-0da6-4ae8-9936-1ebcf9daa38c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.184802] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1055.184802] env[63293]: value = "task-1328201" [ 1055.184802] env[63293]: _type = "Task" [ 1055.184802] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.194471] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328201, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.283311] env[63293]: DEBUG nova.virt.hardware [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1055.283673] env[63293]: DEBUG nova.virt.hardware [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1055.283893] env[63293]: DEBUG nova.virt.hardware [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.284135] env[63293]: DEBUG nova.virt.hardware [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1055.284346] env[63293]: DEBUG nova.virt.hardware [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.284549] env[63293]: DEBUG nova.virt.hardware [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1055.284802] env[63293]: DEBUG nova.virt.hardware [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1055.285042] env[63293]: DEBUG nova.virt.hardware [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1055.285272] env[63293]: DEBUG nova.virt.hardware [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1055.285474] env[63293]: DEBUG nova.virt.hardware [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1055.285702] env[63293]: DEBUG nova.virt.hardware [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.286699] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f381b88a-d1a8-44da-b204-571397ec142c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.296344] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac5a4f20-61c4-4a61-a055-cf057dab5376 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.312047] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Instance VIF info [] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1055.318144] env[63293]: DEBUG oslo.service.loopingcall [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.318457] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1055.318770] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3405581-b386-43ee-8ad0-d2481273782d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.341411] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1055.341411] env[63293]: value = "task-1328202" [ 1055.341411] env[63293]: _type = "Task" [ 1055.341411] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.350903] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328202, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.387063] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.693s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.389587] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.553s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.389859] env[63293]: DEBUG nova.objects.instance [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lazy-loading 'resources' on Instance uuid 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.410329] env[63293]: INFO nova.scheduler.client.report [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Deleted allocations for instance 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50 [ 1055.695062] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328201, 'name': PowerOffVM_Task, 'duration_secs': 0.192303} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.695946] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1055.696705] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ada285e-bef6-4932-acd2-eff72b608443 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.716336] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fab4cb7-a1f6-4b48-9fff-14f958e11003 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.743826] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1055.744134] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cb61798-8ce1-4ebf-9ae2-3609e5f85700 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.751474] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1055.751474] env[63293]: value = "task-1328203" [ 1055.751474] env[63293]: _type = "Task" [ 1055.751474] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.760722] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328203, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.851962] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328202, 'name': CreateVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.919064] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f01861be-4ab3-46a1-8c21-22c8640e5ea3 tempest-ServersV294TestFqdnHostnames-58252378 tempest-ServersV294TestFqdnHostnames-58252378-project-member] Lock "2bf4a8a9-c46d-4512-96b7-b34b1bf69f50" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.125s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.992657] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36963baa-8985-447b-a1c9-7976908ebe08 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.000981] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f2478f-3029-481e-9060-12b1e6443baa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.031335] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d50749-3b70-499a-be2e-a28adb0ec11e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.039507] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a0d161-c716-4ce9-8166-aa60ab5166a6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.053212] env[63293]: DEBUG nova.compute.provider_tree [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.263151] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] VM already powered off {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1056.263381] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1056.263629] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.263780] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.263958] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1056.264229] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d89a86e-933b-47b1-bfa1-4b1cae630ab9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.284829] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1056.285037] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1056.285793] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d675c239-9f83-457d-a1f4-0c9392f5987b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.291891] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1056.291891] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5289e1cd-76bb-5eac-53d0-04c450e8a5a0" [ 1056.291891] env[63293]: _type = "Task" [ 1056.291891] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.299690] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5289e1cd-76bb-5eac-53d0-04c450e8a5a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.354345] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328202, 'name': CreateVM_Task} progress is 15%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.368476] env[63293]: DEBUG oslo_concurrency.lockutils [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.368718] env[63293]: DEBUG oslo_concurrency.lockutils [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.556311] env[63293]: DEBUG nova.scheduler.client.report [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.803706] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5289e1cd-76bb-5eac-53d0-04c450e8a5a0, 'name': SearchDatastore_Task, 'duration_secs': 0.033339} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.804526] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43e3acc1-b4e5-4857-8f1a-1e30b9478f22 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.809834] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1056.809834] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525900cc-5e1f-9360-764e-ec0067fccde7" [ 1056.809834] env[63293]: _type = "Task" [ 1056.809834] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.818643] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525900cc-5e1f-9360-764e-ec0067fccde7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.852871] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328202, 'name': CreateVM_Task} progress is 15%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.873087] env[63293]: DEBUG nova.compute.utils [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1057.061748] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.672s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.088196] env[63293]: INFO nova.scheduler.client.report [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Deleted allocations for instance 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2 [ 1057.320545] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525900cc-5e1f-9360-764e-ec0067fccde7, 'name': SearchDatastore_Task, 'duration_secs': 0.025836} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.320826] env[63293]: DEBUG oslo_concurrency.lockutils [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.321122] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 07418025-09fc-4793-a8ea-f059f42251ea/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk. {{(pid=63293) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1057.321407] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7469cc48-9164-4f4e-a472-8fdbb934d410 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.329067] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1057.329067] env[63293]: value = "task-1328204" [ 1057.329067] env[63293]: _type = "Task" [ 1057.329067] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.337707] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328204, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.353708] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328202, 'name': CreateVM_Task} progress is 15%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.376986] env[63293]: DEBUG oslo_concurrency.lockutils [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.598577] env[63293]: DEBUG oslo_concurrency.lockutils [None req-11a1c580-99c2-43e2-b47e-856190a41f60 tempest-AttachVolumeTestJSON-650225122 tempest-AttachVolumeTestJSON-650225122-project-member] Lock "9937a3c4-c3a6-423b-ac4f-50b7ec5879d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.578s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.840464] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328204, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.854543] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328202, 'name': CreateVM_Task} progress is 15%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.341116] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328204, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.620009} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.341429] env[63293]: INFO nova.virt.vmwareapi.ds_util [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 07418025-09fc-4793-a8ea-f059f42251ea/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk. [ 1058.342347] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a498696-eee3-428c-adca-985acd6dd82a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.356064] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328202, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.377152] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 07418025-09fc-4793-a8ea-f059f42251ea/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1058.377491] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e6a9464-6f25-4108-a635-02748e77b354 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.397252] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1058.397252] env[63293]: value = "task-1328206" [ 1058.397252] env[63293]: _type = "Task" [ 1058.397252] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.406548] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328206, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.444813] env[63293]: DEBUG oslo_concurrency.lockutils [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.445052] env[63293]: DEBUG oslo_concurrency.lockutils [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.445345] env[63293]: INFO nova.compute.manager [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Attaching volume c4a57846-fa87-44ef-a7e0-0cd64f5a6037 to /dev/sdb [ 1058.482461] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8428690c-9441-4ed2-be0b-ed22f3ce8e0b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.493967] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7de160-793c-46db-8fad-b714af494483 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.513834] env[63293]: DEBUG nova.virt.block_device [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating existing volume attachment record: 297a625f-b22f-4cd8-a0cd-d96fd3e49062 {{(pid=63293) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1058.856313] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328202, 'name': CreateVM_Task, 'duration_secs': 3.358855} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.856648] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1058.856930] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.857106] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.857452] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1058.857720] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d52c6637-7480-4016-9def-726b98a8137e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.864068] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1058.864068] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5217be4f-f183-9138-773c-1eddaa79231e" [ 1058.864068] env[63293]: _type = "Task" [ 1058.864068] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.875368] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5217be4f-f183-9138-773c-1eddaa79231e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.910581] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328206, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.376534] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5217be4f-f183-9138-773c-1eddaa79231e, 'name': SearchDatastore_Task, 'duration_secs': 0.031912} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.376869] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.377120] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1059.377357] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.377542] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.377767] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1059.378051] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d39ad9ae-433c-4c84-ac69-e2768b6b708f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.387675] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1059.387872] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1059.388660] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcfdb254-a7b3-4185-9de1-e98d4e81891a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.395343] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1059.395343] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5299f75d-5737-d23e-89a3-136ba70fd96b" [ 1059.395343] env[63293]: _type = "Task" [ 1059.395343] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.408771] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328206, 'name': ReconfigVM_Task, 'duration_secs': 0.738713} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.412815] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 07418025-09fc-4793-a8ea-f059f42251ea/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1059.413384] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5299f75d-5737-d23e-89a3-136ba70fd96b, 'name': SearchDatastore_Task, 'duration_secs': 0.011085} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.414069] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899e5547-408b-4ebe-af4f-ba6146faae88 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.417249] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7868b9a-d324-4c23-a31d-4129afed0512 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.424173] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1059.424173] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ed2602-f65a-f83d-ce76-8a3ab000a1f6" [ 1059.424173] env[63293]: _type = "Task" [ 1059.424173] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.450656] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36d7b8ca-e23c-4d47-91f5-a65445ca08b3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.470054] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ed2602-f65a-f83d-ce76-8a3ab000a1f6, 'name': SearchDatastore_Task, 'duration_secs': 0.011823} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.471530] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.471825] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] b0171e46-ec0b-4fd6-bb04-6f943757b083/b0171e46-ec0b-4fd6-bb04-6f943757b083.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1059.472236] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1059.472236] env[63293]: value = "task-1328210" [ 1059.472236] env[63293]: _type = "Task" [ 1059.472236] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.472472] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7769743a-7f25-4c66-be1d-9ed8a6480a00 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.486415] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328210, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.488056] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1059.488056] env[63293]: value = "task-1328211" [ 1059.488056] env[63293]: _type = "Task" [ 1059.488056] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.496984] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328211, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.985361] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328210, 'name': ReconfigVM_Task, 'duration_secs': 0.351718} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.985653] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1059.986013] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b102a634-8942-4153-b075-8f7d8e633491 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.995424] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1059.995424] env[63293]: value = "task-1328212" [ 1059.995424] env[63293]: _type = "Task" [ 1059.995424] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.999181] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328211, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.008868] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328212, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.501074] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328211, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.507907] env[63293]: DEBUG oslo_vmware.api [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328212, 'name': PowerOnVM_Task, 'duration_secs': 0.411123} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.508225] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1060.511228] env[63293]: DEBUG nova.compute.manager [None req-44b5417a-7a2d-4141-aedf-633f4983a394 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1060.511985] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c1222d-5803-4202-ab19-7fadafea447e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.003485] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328211, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.504522] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328211, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.003441] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328211, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.429644} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.003750] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] b0171e46-ec0b-4fd6-bb04-6f943757b083/b0171e46-ec0b-4fd6-bb04-6f943757b083.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1062.004040] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1062.004349] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63f106e0-7168-4d21-97f6-35eadc42358a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.012548] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1062.012548] env[63293]: value = "task-1328214" [ 1062.012548] env[63293]: _type = "Task" [ 1062.012548] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.022924] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328214, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.279886] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.280338] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.526706] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328214, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081148} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.527153] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1062.528298] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf3cb6a-20e3-4431-9a77-051f768dd232 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.559451] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] b0171e46-ec0b-4fd6-bb04-6f943757b083/b0171e46-ec0b-4fd6-bb04-6f943757b083.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.560158] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3263705-1190-49e1-ae22-eaab78844053 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.585328] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1062.585328] env[63293]: value = "task-1328215" [ 1062.585328] env[63293]: _type = "Task" [ 1062.585328] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.594566] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328215, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.782610] env[63293]: DEBUG nova.compute.manager [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1063.097541] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328215, 'name': ReconfigVM_Task, 'duration_secs': 0.291041} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.097895] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Reconfigured VM instance instance-0000006a to attach disk [datastore1] b0171e46-ec0b-4fd6-bb04-6f943757b083/b0171e46-ec0b-4fd6-bb04-6f943757b083.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.098529] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8bf3553-7dc1-4780-bf50-20d9d53e7291 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.106440] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1063.106440] env[63293]: value = "task-1328217" [ 1063.106440] env[63293]: _type = "Task" [ 1063.106440] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.116279] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328217, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.304377] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.304658] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.306186] env[63293]: INFO nova.compute.claims [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1063.505376] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquiring lock "4969456a-be05-42d1-a4d5-44dc537f5a21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.505376] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Lock "4969456a-be05-42d1-a4d5-44dc537f5a21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.623049] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328217, 'name': Rename_Task, 'duration_secs': 0.155171} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.623690] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1063.625741] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1759762b-3371-4767-ac3b-41d7125a9ee3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.633742] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1063.633742] env[63293]: value = "task-1328219" [ 1063.633742] env[63293]: _type = "Task" [ 1063.633742] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.648814] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328219, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.007703] env[63293]: DEBUG nova.compute.manager [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1064.129960] env[63293]: DEBUG oslo_concurrency.lockutils [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.130297] env[63293]: DEBUG oslo_concurrency.lockutils [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.130537] env[63293]: INFO nova.compute.manager [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Shelving [ 1064.145074] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328219, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.440154] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430a1a23-9768-4118-bd64-90d1c87b3fcc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.448488] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484fb231-233f-46d5-ac4c-554df4f8f5cf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.479116] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0e890a-2d18-4cc0-9ed8-9fa70355f66a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.487145] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-922bc657-bc61-4c6a-97fc-ec9253286df8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.502307] env[63293]: DEBUG nova.compute.provider_tree [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.526417] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.641379] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1064.641632] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4a2fb1f-f3b8-4680-8e6d-15b7af59e12e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.646730] env[63293]: DEBUG oslo_vmware.api [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328219, 'name': PowerOnVM_Task, 'duration_secs': 0.905887} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.646980] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1064.647257] env[63293]: DEBUG nova.compute.manager [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1064.648890] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ba9ee2-754e-485b-8adf-d44f7e1b90bb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.652381] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1064.652381] env[63293]: value = "task-1328220" [ 1064.652381] env[63293]: _type = "Task" [ 1064.652381] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.664859] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328220, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.006207] env[63293]: DEBUG nova.scheduler.client.report [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.070323] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Volume attach. Driver type: vmdk {{(pid=63293) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1065.070624] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283845', 'volume_id': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'name': 'volume-c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bea1ea16-c224-4a0e-86ed-c2aaf920d3e1', 'attached_at': '', 'detached_at': '', 'volume_id': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'serial': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1065.071608] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52095dd-8b67-4465-a9d6-d85a7fa9dc70 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.099733] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18fa5bc8-a27c-45d7-90d1-2052d41556b2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.124402] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] volume-c4a57846-fa87-44ef-a7e0-0cd64f5a6037/volume-c4a57846-fa87-44ef-a7e0-0cd64f5a6037.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.124695] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf763c97-f57c-41b1-8876-56a087a0ace9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.144111] env[63293]: DEBUG oslo_vmware.api [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1065.144111] env[63293]: value = "task-1328221" [ 1065.144111] env[63293]: _type = "Task" [ 1065.144111] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.152547] env[63293]: DEBUG oslo_vmware.api [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328221, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.160263] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328220, 'name': PowerOffVM_Task, 'duration_secs': 0.328181} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.160509] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1065.164791] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7d1aeb-d742-4df8-ab92-4283b5bec771 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.168830] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.184310] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7268d53d-6e29-4a54-b5e3-74fdbc7fa906 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.512074] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.207s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.512788] env[63293]: DEBUG nova.compute.manager [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1065.515408] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.989s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.517688] env[63293]: INFO nova.compute.claims [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1065.655063] env[63293]: DEBUG oslo_vmware.api [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328221, 'name': ReconfigVM_Task, 'duration_secs': 0.352916} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.655170] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfigured VM instance instance-00000064 to attach disk [datastore1] volume-c4a57846-fa87-44ef-a7e0-0cd64f5a6037/volume-c4a57846-fa87-44ef-a7e0-0cd64f5a6037.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1065.659861] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d1d0bb1-d9cc-493e-a304-1da7ac1b5328 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.676551] env[63293]: DEBUG oslo_vmware.api [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1065.676551] env[63293]: value = "task-1328222" [ 1065.676551] env[63293]: _type = "Task" [ 1065.676551] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.687200] env[63293]: DEBUG oslo_vmware.api [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328222, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.694714] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Creating Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1065.695015] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-12fb2a12-ae6e-4b6e-9766-776440ef0f27 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.703300] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1065.703300] env[63293]: value = "task-1328223" [ 1065.703300] env[63293]: _type = "Task" [ 1065.703300] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.711972] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328223, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.022077] env[63293]: DEBUG nova.compute.utils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1066.026148] env[63293]: DEBUG nova.compute.manager [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1066.026379] env[63293]: DEBUG nova.network.neutron [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1066.082480] env[63293]: DEBUG nova.policy [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '53698361c41947a38f1b393e0d5290c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8adfd01c54641fb8bee8cd534d50569', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 1066.187458] env[63293]: DEBUG oslo_vmware.api [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328222, 'name': ReconfigVM_Task, 'duration_secs': 0.181128} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.190138] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283845', 'volume_id': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'name': 'volume-c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bea1ea16-c224-4a0e-86ed-c2aaf920d3e1', 'attached_at': '', 'detached_at': '', 'volume_id': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'serial': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1066.213571] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328223, 'name': CreateSnapshot_Task, 'duration_secs': 0.426243} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.214261] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Created Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1066.214887] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f8e151-ae78-448e-bfbf-d07b2d10f9ac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.355075] env[63293]: DEBUG nova.network.neutron [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Successfully created port: eef289b9-0d13-4ce5-be81-a5247187e457 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1066.526948] env[63293]: DEBUG nova.compute.manager [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1066.593731] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "b0171e46-ec0b-4fd6-bb04-6f943757b083" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.594141] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "b0171e46-ec0b-4fd6-bb04-6f943757b083" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.594622] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "b0171e46-ec0b-4fd6-bb04-6f943757b083-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.594622] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "b0171e46-ec0b-4fd6-bb04-6f943757b083-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.594622] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "b0171e46-ec0b-4fd6-bb04-6f943757b083-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.596836] env[63293]: INFO nova.compute.manager [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Terminating instance [ 1066.600780] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "refresh_cache-b0171e46-ec0b-4fd6-bb04-6f943757b083" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.600971] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquired lock "refresh_cache-b0171e46-ec0b-4fd6-bb04-6f943757b083" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.601242] env[63293]: DEBUG nova.network.neutron [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1066.651080] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2179aaf-c7a7-4de3-858f-8b61245ae147 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.659895] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43b8b79-335c-4cfe-b6fe-364b93819160 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.699939] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9270da-00de-4835-b95b-c42c33293ad2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.710304] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644a205d-02a5-4bd4-8ff4-102f059317f1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.734492] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Creating linked-clone VM from snapshot {{(pid=63293) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1066.735270] env[63293]: DEBUG nova.compute.provider_tree [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.740584] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bede59f8-49f4-4b42-ae4b-0ff8e55365b2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.753068] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1066.753068] env[63293]: value = "task-1328224" [ 1066.753068] env[63293]: _type = "Task" [ 1066.753068] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.762796] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328224, 'name': CloneVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.871889] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.871889] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Cleaning up deleted instances with incomplete migration {{(pid=63293) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1067.122216] env[63293]: DEBUG nova.network.neutron [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1067.184951] env[63293]: DEBUG nova.network.neutron [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.242118] env[63293]: DEBUG nova.scheduler.client.report [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.246050] env[63293]: DEBUG nova.objects.instance [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'flavor' on Instance uuid bea1ea16-c224-4a0e-86ed-c2aaf920d3e1 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.266411] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328224, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.541330] env[63293]: DEBUG nova.compute.manager [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1067.573181] env[63293]: DEBUG nova.virt.hardware [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1067.573666] env[63293]: DEBUG nova.virt.hardware [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1067.573930] env[63293]: DEBUG nova.virt.hardware [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1067.574222] env[63293]: DEBUG nova.virt.hardware [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1067.574497] env[63293]: DEBUG nova.virt.hardware [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1067.574745] env[63293]: DEBUG nova.virt.hardware [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1067.575063] env[63293]: DEBUG nova.virt.hardware [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1067.575328] env[63293]: DEBUG nova.virt.hardware [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1067.575593] env[63293]: DEBUG nova.virt.hardware [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1067.575879] env[63293]: DEBUG nova.virt.hardware [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1067.576187] env[63293]: DEBUG nova.virt.hardware [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1067.577178] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf37e00-ea8d-45d9-a903-f5bd6c8b40c5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.586716] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0e1982-1b49-40fb-b6da-028443057bd4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.693022] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Releasing lock "refresh_cache-b0171e46-ec0b-4fd6-bb04-6f943757b083" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.693022] env[63293]: DEBUG nova.compute.manager [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1067.693022] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1067.693022] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69b675e-e722-4645-8091-a9fdd7d67596 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.702515] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1067.703282] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02120b09-1f02-4641-9130-26d23b60a900 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.715063] env[63293]: DEBUG oslo_vmware.api [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1067.715063] env[63293]: value = "task-1328225" [ 1067.715063] env[63293]: _type = "Task" [ 1067.715063] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.725493] env[63293]: DEBUG oslo_vmware.api [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328225, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.751023] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.233s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.751023] env[63293]: DEBUG nova.compute.manager [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1067.752246] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.583s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.752624] env[63293]: DEBUG nova.objects.instance [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63293) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1067.765779] env[63293]: DEBUG oslo_concurrency.lockutils [None req-49158685-78f8-4327-8574-d72e0d3799b6 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.321s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.779128] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328224, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.865712] env[63293]: DEBUG nova.compute.manager [req-f6737e37-56c6-4ef5-bb4b-b94ad9f5bec9 req-9a390478-64b3-4f90-ba54-83ad25064209 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Received event network-vif-plugged-eef289b9-0d13-4ce5-be81-a5247187e457 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.865712] env[63293]: DEBUG oslo_concurrency.lockutils [req-f6737e37-56c6-4ef5-bb4b-b94ad9f5bec9 req-9a390478-64b3-4f90-ba54-83ad25064209 service nova] Acquiring lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.865712] env[63293]: DEBUG oslo_concurrency.lockutils [req-f6737e37-56c6-4ef5-bb4b-b94ad9f5bec9 req-9a390478-64b3-4f90-ba54-83ad25064209 service nova] Lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.865712] env[63293]: DEBUG oslo_concurrency.lockutils [req-f6737e37-56c6-4ef5-bb4b-b94ad9f5bec9 req-9a390478-64b3-4f90-ba54-83ad25064209 service nova] Lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.865712] env[63293]: DEBUG nova.compute.manager [req-f6737e37-56c6-4ef5-bb4b-b94ad9f5bec9 req-9a390478-64b3-4f90-ba54-83ad25064209 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] No waiting events found dispatching network-vif-plugged-eef289b9-0d13-4ce5-be81-a5247187e457 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1067.865712] env[63293]: WARNING nova.compute.manager [req-f6737e37-56c6-4ef5-bb4b-b94ad9f5bec9 req-9a390478-64b3-4f90-ba54-83ad25064209 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Received unexpected event network-vif-plugged-eef289b9-0d13-4ce5-be81-a5247187e457 for instance with vm_state building and task_state spawning. [ 1068.023840] env[63293]: DEBUG nova.network.neutron [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Successfully updated port: eef289b9-0d13-4ce5-be81-a5247187e457 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1068.224112] env[63293]: DEBUG oslo_vmware.api [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328225, 'name': PowerOffVM_Task, 'duration_secs': 0.228438} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.224516] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1068.224626] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1068.224902] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a07511af-3cdc-4547-b754-ac3120b3f2e5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.254539] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1068.254923] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1068.255066] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Deleting the datastore file [datastore1] b0171e46-ec0b-4fd6-bb04-6f943757b083 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1068.255651] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99ae6ff6-41a6-4a21-bf17-ac5c389d5173 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.258936] env[63293]: DEBUG nova.compute.utils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1068.264215] env[63293]: DEBUG nova.compute.manager [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1068.264400] env[63293]: DEBUG nova.network.neutron [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1068.279906] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4520a5ae-25dc-439b-a05d-4fc1a9b30f0e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.527s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.284483] env[63293]: DEBUG oslo_vmware.api [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1068.284483] env[63293]: value = "task-1328227" [ 1068.284483] env[63293]: _type = "Task" [ 1068.284483] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.294278] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328224, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.300465] env[63293]: DEBUG oslo_vmware.api [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328227, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.354341] env[63293]: DEBUG nova.policy [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1b75e640dc24e2eacdd5087545e2735', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '64648e6fe7aa4724aadb76aac933fc2f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 1068.526541] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.526952] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.527194] env[63293]: DEBUG nova.network.neutron [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1068.661575] env[63293]: DEBUG nova.network.neutron [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Successfully created port: 58d85ce7-565c-471e-abde-4f572863d75a {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1068.764733] env[63293]: DEBUG nova.compute.manager [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1068.773559] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328224, 'name': CloneVM_Task, 'duration_secs': 1.828208} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.773883] env[63293]: INFO nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Created linked-clone VM from snapshot [ 1068.775147] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75bcf17-acbb-4804-8435-a0030d28f709 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.784101] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Uploading image e4ccc7b7-9af6-4851-89da-fd23cc904ad3 {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1068.796225] env[63293]: DEBUG oslo_vmware.api [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328227, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137423} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.796461] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.796577] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1068.796756] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1068.796927] env[63293]: INFO nova.compute.manager [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1068.797182] env[63293]: DEBUG oslo.service.loopingcall [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1068.797446] env[63293]: DEBUG nova.compute.manager [-] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1068.797542] env[63293]: DEBUG nova.network.neutron [-] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1068.817709] env[63293]: DEBUG oslo_vmware.rw_handles [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1068.817709] env[63293]: value = "vm-283847" [ 1068.817709] env[63293]: _type = "VirtualMachine" [ 1068.817709] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1068.818014] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0872cbc5-1d1f-4c1d-b061-8da723b1d53f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.823535] env[63293]: DEBUG nova.network.neutron [-] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1068.826786] env[63293]: DEBUG oslo_vmware.rw_handles [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lease: (returnval){ [ 1068.826786] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528f7d3d-f72d-567a-6f09-e54d66af4795" [ 1068.826786] env[63293]: _type = "HttpNfcLease" [ 1068.826786] env[63293]: } obtained for exporting VM: (result){ [ 1068.826786] env[63293]: value = "vm-283847" [ 1068.826786] env[63293]: _type = "VirtualMachine" [ 1068.826786] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1068.827244] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the lease: (returnval){ [ 1068.827244] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528f7d3d-f72d-567a-6f09-e54d66af4795" [ 1068.827244] env[63293]: _type = "HttpNfcLease" [ 1068.827244] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1068.834706] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1068.834706] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528f7d3d-f72d-567a-6f09-e54d66af4795" [ 1068.834706] env[63293]: _type = "HttpNfcLease" [ 1068.834706] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1068.928080] env[63293]: DEBUG nova.compute.manager [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Stashing vm_state: active {{(pid=63293) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1069.059337] env[63293]: DEBUG nova.network.neutron [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1069.171955] env[63293]: DEBUG nova.network.neutron [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Updating instance_info_cache with network_info: [{"id": "eef289b9-0d13-4ce5-be81-a5247187e457", "address": "fa:16:3e:e5:f5:fb", "network": {"id": "69a5064b-e3a0-4ef6-b722-09fbb9b50ca8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1786908568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b8adfd01c54641fb8bee8cd534d50569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef289b9-0d", "ovs_interfaceid": "eef289b9-0d13-4ce5-be81-a5247187e457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.326094] env[63293]: DEBUG nova.network.neutron [-] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.336698] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1069.336698] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528f7d3d-f72d-567a-6f09-e54d66af4795" [ 1069.336698] env[63293]: _type = "HttpNfcLease" [ 1069.336698] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1069.336698] env[63293]: DEBUG oslo_vmware.rw_handles [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1069.336698] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]528f7d3d-f72d-567a-6f09-e54d66af4795" [ 1069.336698] env[63293]: _type = "HttpNfcLease" [ 1069.336698] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1069.337503] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc095b27-51c3-4fce-8627-abce71df1b4b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.346559] env[63293]: DEBUG oslo_vmware.rw_handles [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525b9c46-0d4b-469c-eb2a-d9be683600bf/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1069.346813] env[63293]: DEBUG oslo_vmware.rw_handles [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525b9c46-0d4b-469c-eb2a-d9be683600bf/disk-0.vmdk for reading. {{(pid=63293) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1069.405785] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.440032] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-02dce99f-3b60-44d9-b76c-5cfd14505589 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.455534] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.455960] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.674455] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Releasing lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.674838] env[63293]: DEBUG nova.compute.manager [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Instance network_info: |[{"id": "eef289b9-0d13-4ce5-be81-a5247187e457", "address": "fa:16:3e:e5:f5:fb", "network": {"id": "69a5064b-e3a0-4ef6-b722-09fbb9b50ca8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1786908568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b8adfd01c54641fb8bee8cd534d50569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef289b9-0d", "ovs_interfaceid": "eef289b9-0d13-4ce5-be81-a5247187e457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1069.675344] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e5:f5:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bff6c3a1-cc80-46ca-86c0-6dbb029edddb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eef289b9-0d13-4ce5-be81-a5247187e457', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1069.683751] env[63293]: DEBUG oslo.service.loopingcall [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.683984] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1069.684630] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a34e301-37da-4d09-a283-454e6bd297d4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.709888] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1069.709888] env[63293]: value = "task-1328229" [ 1069.709888] env[63293]: _type = "Task" [ 1069.709888] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.721591] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328229, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.774703] env[63293]: DEBUG nova.compute.manager [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1069.797167] env[63293]: DEBUG nova.virt.hardware [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1069.797465] env[63293]: DEBUG nova.virt.hardware [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1069.797643] env[63293]: DEBUG nova.virt.hardware [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1069.797833] env[63293]: DEBUG nova.virt.hardware [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1069.797983] env[63293]: DEBUG nova.virt.hardware [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1069.798147] env[63293]: DEBUG nova.virt.hardware [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1069.798360] env[63293]: DEBUG nova.virt.hardware [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1069.798529] env[63293]: DEBUG nova.virt.hardware [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1069.798738] env[63293]: DEBUG nova.virt.hardware [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1069.798927] env[63293]: DEBUG nova.virt.hardware [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1069.799172] env[63293]: DEBUG nova.virt.hardware [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1069.800067] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55bc3781-b2dc-4cad-ac09-0590d3af4b64 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.809010] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8c25dd-9d29-4274-b721-6eadd556b1a4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.828621] env[63293]: INFO nova.compute.manager [-] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Took 1.03 seconds to deallocate network for instance. [ 1069.871355] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.871590] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.871787] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63293) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1069.872611] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.893671] env[63293]: DEBUG nova.compute.manager [req-57e17021-b37c-410d-b6d6-8e82bae7888c req-ddbaead9-ca59-4fd5-91e3-b56812b05c13 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Received event network-changed-eef289b9-0d13-4ce5-be81-a5247187e457 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.893671] env[63293]: DEBUG nova.compute.manager [req-57e17021-b37c-410d-b6d6-8e82bae7888c req-ddbaead9-ca59-4fd5-91e3-b56812b05c13 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Refreshing instance network info cache due to event network-changed-eef289b9-0d13-4ce5-be81-a5247187e457. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1069.893671] env[63293]: DEBUG oslo_concurrency.lockutils [req-57e17021-b37c-410d-b6d6-8e82bae7888c req-ddbaead9-ca59-4fd5-91e3-b56812b05c13 service nova] Acquiring lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.893671] env[63293]: DEBUG oslo_concurrency.lockutils [req-57e17021-b37c-410d-b6d6-8e82bae7888c req-ddbaead9-ca59-4fd5-91e3-b56812b05c13 service nova] Acquired lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.893671] env[63293]: DEBUG nova.network.neutron [req-57e17021-b37c-410d-b6d6-8e82bae7888c req-ddbaead9-ca59-4fd5-91e3-b56812b05c13 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Refreshing network info cache for port eef289b9-0d13-4ce5-be81-a5247187e457 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1069.961183] env[63293]: INFO nova.compute.claims [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1070.201747] env[63293]: DEBUG nova.compute.manager [req-2a17c0df-7630-4376-a42f-08aad1f77c22 req-6cfe917b-766f-4879-8e7f-7ccf481b36bb service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Received event network-vif-plugged-58d85ce7-565c-471e-abde-4f572863d75a {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1070.201976] env[63293]: DEBUG oslo_concurrency.lockutils [req-2a17c0df-7630-4376-a42f-08aad1f77c22 req-6cfe917b-766f-4879-8e7f-7ccf481b36bb service nova] Acquiring lock "4969456a-be05-42d1-a4d5-44dc537f5a21-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.202244] env[63293]: DEBUG oslo_concurrency.lockutils [req-2a17c0df-7630-4376-a42f-08aad1f77c22 req-6cfe917b-766f-4879-8e7f-7ccf481b36bb service nova] Lock "4969456a-be05-42d1-a4d5-44dc537f5a21-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.202477] env[63293]: DEBUG oslo_concurrency.lockutils [req-2a17c0df-7630-4376-a42f-08aad1f77c22 req-6cfe917b-766f-4879-8e7f-7ccf481b36bb service nova] Lock "4969456a-be05-42d1-a4d5-44dc537f5a21-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.202970] env[63293]: DEBUG nova.compute.manager [req-2a17c0df-7630-4376-a42f-08aad1f77c22 req-6cfe917b-766f-4879-8e7f-7ccf481b36bb service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] No waiting events found dispatching network-vif-plugged-58d85ce7-565c-471e-abde-4f572863d75a {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1070.203399] env[63293]: WARNING nova.compute.manager [req-2a17c0df-7630-4376-a42f-08aad1f77c22 req-6cfe917b-766f-4879-8e7f-7ccf481b36bb service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Received unexpected event network-vif-plugged-58d85ce7-565c-471e-abde-4f572863d75a for instance with vm_state building and task_state spawning. [ 1070.222151] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328229, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.291949] env[63293]: DEBUG nova.network.neutron [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Successfully updated port: 58d85ce7-565c-471e-abde-4f572863d75a {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1070.335188] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.375376] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.467159] env[63293]: INFO nova.compute.resource_tracker [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating resource usage from migration eff19c51-22a9-426e-9f8b-a1e793420d8e [ 1070.619831] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2826104f-f4f3-4805-a552-40e89f5a21ac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.628346] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e55d483-7937-4333-bab4-3de0602a48f2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.667773] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d47ea8-5c0a-475e-9ffa-87bc1ff58e18 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.676491] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba65480b-a392-4a78-a6f6-5f121cea2e85 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.692853] env[63293]: DEBUG nova.compute.provider_tree [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.722711] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328229, 'name': CreateVM_Task, 'duration_secs': 0.735124} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.723081] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1070.723806] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.724070] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.724662] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1070.724905] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e5d697b-aa31-4c59-b419-2dbdcba6b36a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.730119] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1070.730119] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52cf98d1-7be9-6a81-19f4-3cc2a40af5e3" [ 1070.730119] env[63293]: _type = "Task" [ 1070.730119] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.734123] env[63293]: DEBUG nova.network.neutron [req-57e17021-b37c-410d-b6d6-8e82bae7888c req-ddbaead9-ca59-4fd5-91e3-b56812b05c13 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Updated VIF entry in instance network info cache for port eef289b9-0d13-4ce5-be81-a5247187e457. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1070.734662] env[63293]: DEBUG nova.network.neutron [req-57e17021-b37c-410d-b6d6-8e82bae7888c req-ddbaead9-ca59-4fd5-91e3-b56812b05c13 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Updating instance_info_cache with network_info: [{"id": "eef289b9-0d13-4ce5-be81-a5247187e457", "address": "fa:16:3e:e5:f5:fb", "network": {"id": "69a5064b-e3a0-4ef6-b722-09fbb9b50ca8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1786908568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b8adfd01c54641fb8bee8cd534d50569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef289b9-0d", "ovs_interfaceid": "eef289b9-0d13-4ce5-be81-a5247187e457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.742397] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52cf98d1-7be9-6a81-19f4-3cc2a40af5e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.795084] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquiring lock "refresh_cache-4969456a-be05-42d1-a4d5-44dc537f5a21" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.795084] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquired lock "refresh_cache-4969456a-be05-42d1-a4d5-44dc537f5a21" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.795305] env[63293]: DEBUG nova.network.neutron [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1071.195942] env[63293]: DEBUG nova.scheduler.client.report [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.237689] env[63293]: DEBUG oslo_concurrency.lockutils [req-57e17021-b37c-410d-b6d6-8e82bae7888c req-ddbaead9-ca59-4fd5-91e3-b56812b05c13 service nova] Releasing lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.242816] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52cf98d1-7be9-6a81-19f4-3cc2a40af5e3, 'name': SearchDatastore_Task, 'duration_secs': 0.013621} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.243315] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.243641] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1071.244030] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.244416] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.244646] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1071.245232] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d36c8383-7be3-4d94-8fd8-ff9ff7cb25e9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.257108] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1071.257363] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1071.258177] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20b0ec78-da8d-47f0-8c48-ccd2dd2cec82 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.264383] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1071.264383] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c39951-c294-72b0-5ee5-89754e0b0b35" [ 1071.264383] env[63293]: _type = "Task" [ 1071.264383] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.272444] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c39951-c294-72b0-5ee5-89754e0b0b35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.327927] env[63293]: DEBUG nova.network.neutron [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1071.464438] env[63293]: DEBUG nova.network.neutron [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Updating instance_info_cache with network_info: [{"id": "58d85ce7-565c-471e-abde-4f572863d75a", "address": "fa:16:3e:b0:33:d9", "network": {"id": "106fb3f0-f7e3-4df9-9529-d2d1c7443e9b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-2003024949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64648e6fe7aa4724aadb76aac933fc2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58d85ce7-56", "ovs_interfaceid": "58d85ce7-565c-471e-abde-4f572863d75a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.701060] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.245s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.701329] env[63293]: INFO nova.compute.manager [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Migrating [ 1071.708173] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.373s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.708433] env[63293]: DEBUG nova.objects.instance [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lazy-loading 'resources' on Instance uuid b0171e46-ec0b-4fd6-bb04-6f943757b083 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.777483] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c39951-c294-72b0-5ee5-89754e0b0b35, 'name': SearchDatastore_Task, 'duration_secs': 0.020731} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.782531] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a96c547-ee83-435e-99a7-846f42aa9ef5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.790217] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1071.790217] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ff742f-1abc-145f-b5c4-990a0c07f086" [ 1071.790217] env[63293]: _type = "Task" [ 1071.790217] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.808440] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52ff742f-1abc-145f-b5c4-990a0c07f086, 'name': SearchDatastore_Task, 'duration_secs': 0.012186} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.808440] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.808440] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] ba8404c7-b43e-4776-941e-4d9bd8fb4c9d/ba8404c7-b43e-4776-941e-4d9bd8fb4c9d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1071.808440] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-868b0d1a-94f7-4788-9fb1-b4baa7ea6b97 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.816883] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1071.816883] env[63293]: value = "task-1328230" [ 1071.816883] env[63293]: _type = "Task" [ 1071.816883] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.826096] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328230, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.879460] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f52123-1363-4bb0-8ffa-e59a7be8debc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.887825] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184d6ad3-932a-4c05-b692-3b6d6f8353b5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.921528] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b675ffe6-b725-4f8c-97fc-efc8f15493ad {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.925260] env[63293]: DEBUG nova.compute.manager [req-a19452e0-42f2-46ff-95e8-691f05197623 req-47007ef1-2577-4cfd-a50e-c353b2fd43a0 service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Received event network-changed-58d85ce7-565c-471e-abde-4f572863d75a {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1071.925455] env[63293]: DEBUG nova.compute.manager [req-a19452e0-42f2-46ff-95e8-691f05197623 req-47007ef1-2577-4cfd-a50e-c353b2fd43a0 service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Refreshing instance network info cache due to event network-changed-58d85ce7-565c-471e-abde-4f572863d75a. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1071.925647] env[63293]: DEBUG oslo_concurrency.lockutils [req-a19452e0-42f2-46ff-95e8-691f05197623 req-47007ef1-2577-4cfd-a50e-c353b2fd43a0 service nova] Acquiring lock "refresh_cache-4969456a-be05-42d1-a4d5-44dc537f5a21" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.936409] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d55d7ea-a335-40eb-933e-771ced5b47b1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.954100] env[63293]: DEBUG nova.compute.provider_tree [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.966474] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Releasing lock "refresh_cache-4969456a-be05-42d1-a4d5-44dc537f5a21" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.966815] env[63293]: DEBUG nova.compute.manager [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Instance network_info: |[{"id": "58d85ce7-565c-471e-abde-4f572863d75a", "address": "fa:16:3e:b0:33:d9", "network": {"id": "106fb3f0-f7e3-4df9-9529-d2d1c7443e9b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-2003024949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64648e6fe7aa4724aadb76aac933fc2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58d85ce7-56", "ovs_interfaceid": "58d85ce7-565c-471e-abde-4f572863d75a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1071.967169] env[63293]: DEBUG oslo_concurrency.lockutils [req-a19452e0-42f2-46ff-95e8-691f05197623 req-47007ef1-2577-4cfd-a50e-c353b2fd43a0 service nova] Acquired lock "refresh_cache-4969456a-be05-42d1-a4d5-44dc537f5a21" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.967358] env[63293]: DEBUG nova.network.neutron [req-a19452e0-42f2-46ff-95e8-691f05197623 req-47007ef1-2577-4cfd-a50e-c353b2fd43a0 service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Refreshing network info cache for port 58d85ce7-565c-471e-abde-4f572863d75a {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1071.968541] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:33:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '31ac3fea-ebf4-4bed-bf70-1eaecdf71280', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58d85ce7-565c-471e-abde-4f572863d75a', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1071.976168] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Creating folder: Project (64648e6fe7aa4724aadb76aac933fc2f). Parent ref: group-v283678. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1071.977383] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-229aec0c-579d-4ea7-af86-81f6edee629c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.991103] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Created folder: Project (64648e6fe7aa4724aadb76aac933fc2f) in parent group-v283678. [ 1071.991398] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Creating folder: Instances. Parent ref: group-v283849. {{(pid=63293) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1071.991674] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f33b5c6-d730-4692-bd27-36ecccb9e75d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.002371] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Created folder: Instances in parent group-v283849. [ 1072.002672] env[63293]: DEBUG oslo.service.loopingcall [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.002895] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1072.003172] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f137fe95-6ea6-46ab-829f-9c1f4386dfae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.026670] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1072.026670] env[63293]: value = "task-1328233" [ 1072.026670] env[63293]: _type = "Task" [ 1072.026670] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.042167] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328233, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.220506] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.220748] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.220936] env[63293]: DEBUG nova.network.neutron [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1072.328941] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328230, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.457892] env[63293]: DEBUG nova.scheduler.client.report [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1072.538883] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328233, 'name': CreateVM_Task} progress is 99%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.690054] env[63293]: DEBUG nova.network.neutron [req-a19452e0-42f2-46ff-95e8-691f05197623 req-47007ef1-2577-4cfd-a50e-c353b2fd43a0 service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Updated VIF entry in instance network info cache for port 58d85ce7-565c-471e-abde-4f572863d75a. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1072.690424] env[63293]: DEBUG nova.network.neutron [req-a19452e0-42f2-46ff-95e8-691f05197623 req-47007ef1-2577-4cfd-a50e-c353b2fd43a0 service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Updating instance_info_cache with network_info: [{"id": "58d85ce7-565c-471e-abde-4f572863d75a", "address": "fa:16:3e:b0:33:d9", "network": {"id": "106fb3f0-f7e3-4df9-9529-d2d1c7443e9b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-2003024949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64648e6fe7aa4724aadb76aac933fc2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58d85ce7-56", "ovs_interfaceid": "58d85ce7-565c-471e-abde-4f572863d75a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.831150] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328230, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544354} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.831470] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] ba8404c7-b43e-4776-941e-4d9bd8fb4c9d/ba8404c7-b43e-4776-941e-4d9bd8fb4c9d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1072.831711] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.831999] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2b62087-4b1e-4ebc-b3e2-b83db600d05a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.840249] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1072.840249] env[63293]: value = "task-1328234" [ 1072.840249] env[63293]: _type = "Task" [ 1072.840249] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.851646] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328234, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.963286] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.255s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.966371] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.591s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.966550] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.966709] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63293) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1072.967700] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3115c89a-3add-4762-8d18-6e1769b35586 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.976845] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70bf63de-c6d5-4a5c-806c-7792465c8eb0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.996302] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0794c2-108d-4462-8394-de56c94c70df {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.004711] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4228876-2493-4e47-aeb4-2541fe6e7f3f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.035981] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180583MB free_disk=132GB free_vcpus=48 pci_devices=None {{(pid=63293) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1073.036250] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.036371] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.038885] env[63293]: INFO nova.scheduler.client.report [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Deleted allocations for instance b0171e46-ec0b-4fd6-bb04-6f943757b083 [ 1073.049017] env[63293]: DEBUG nova.network.neutron [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance_info_cache with network_info: [{"id": "c589de77-25fc-43f7-9480-e6c97bbda93d", "address": "fa:16:3e:4a:ef:d2", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc589de77-25", "ovs_interfaceid": "c589de77-25fc-43f7-9480-e6c97bbda93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.055169] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328233, 'name': CreateVM_Task, 'duration_secs': 0.583926} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.055676] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1073.056591] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.056924] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.057364] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1073.058413] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52a7de48-cb83-4ead-93b2-97b12e3d53bf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.066165] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for the task: (returnval){ [ 1073.066165] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522625b8-8509-3709-2323-bb4a61aca173" [ 1073.066165] env[63293]: _type = "Task" [ 1073.066165] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.077370] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522625b8-8509-3709-2323-bb4a61aca173, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.193404] env[63293]: DEBUG oslo_concurrency.lockutils [req-a19452e0-42f2-46ff-95e8-691f05197623 req-47007ef1-2577-4cfd-a50e-c353b2fd43a0 service nova] Releasing lock "refresh_cache-4969456a-be05-42d1-a4d5-44dc537f5a21" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.351281] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328234, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069357} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.351575] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1073.352404] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8811255-d505-4b29-b940-f0e6cbc03d57 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.374735] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] ba8404c7-b43e-4776-941e-4d9bd8fb4c9d/ba8404c7-b43e-4776-941e-4d9bd8fb4c9d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.375029] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4aa46b6-9bc9-4c32-8447-60e8eb5efd37 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.397758] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1073.397758] env[63293]: value = "task-1328235" [ 1073.397758] env[63293]: _type = "Task" [ 1073.397758] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.412998] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328235, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.550668] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4cd03805-c7a8-49da-8772-384cbfac39fe tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "b0171e46-ec0b-4fd6-bb04-6f943757b083" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.956s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.556133] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.579816] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]522625b8-8509-3709-2323-bb4a61aca173, 'name': SearchDatastore_Task, 'duration_secs': 0.013256} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.580394] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.580645] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1073.580880] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.581042] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.581235] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1073.581517] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a2a074c-d617-403f-8f4a-e58b642577bc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.590687] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1073.590879] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1073.591649] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df62c833-81a8-4eb7-9426-95b7a473b8f4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.597786] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for the task: (returnval){ [ 1073.597786] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529e809c-ad01-482f-8a7b-9038039e7eca" [ 1073.597786] env[63293]: _type = "Task" [ 1073.597786] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.606290] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529e809c-ad01-482f-8a7b-9038039e7eca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.908164] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328235, 'name': ReconfigVM_Task, 'duration_secs': 0.382686} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.908461] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Reconfigured VM instance instance-0000006b to attach disk [datastore1] ba8404c7-b43e-4776-941e-4d9bd8fb4c9d/ba8404c7-b43e-4776-941e-4d9bd8fb4c9d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.909117] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abdaf71f-b918-4a1b-97a7-57ada0557a34 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.918024] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1073.918024] env[63293]: value = "task-1328236" [ 1073.918024] env[63293]: _type = "Task" [ 1073.918024] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.927555] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328236, 'name': Rename_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.050969] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Applying migration context for instance bea1ea16-c224-4a0e-86ed-c2aaf920d3e1 as it has an incoming, in-progress migration eff19c51-22a9-426e-9f8b-a1e793420d8e. Migration status is pre-migrating {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1074.052185] env[63293]: INFO nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating resource usage from migration eff19c51-22a9-426e-9f8b-a1e793420d8e [ 1074.071735] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.071735] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 97da107a-2e15-4da2-9197-7f85d960ff27 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.071993] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 07418025-09fc-4793-a8ea-f059f42251ea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.071993] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 0041acd3-de76-4745-a1ab-c6563f774f33 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.072082] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance ba8404c7-b43e-4776-941e-4d9bd8fb4c9d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.072182] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 4969456a-be05-42d1-a4d5-44dc537f5a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.072302] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Migration eff19c51-22a9-426e-9f8b-a1e793420d8e is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1074.072421] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance bea1ea16-c224-4a0e-86ed-c2aaf920d3e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.072616] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1074.072752] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1074.112283] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529e809c-ad01-482f-8a7b-9038039e7eca, 'name': SearchDatastore_Task, 'duration_secs': 0.011876} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.113134] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9b13194-248d-41ea-b4eb-c5612bcac49c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.122253] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for the task: (returnval){ [ 1074.122253] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52035aa5-04e5-e7c2-f347-ca019b0d03e1" [ 1074.122253] env[63293]: _type = "Task" [ 1074.122253] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.132542] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52035aa5-04e5-e7c2-f347-ca019b0d03e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.183546] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bfdd2b5-451c-460e-b29e-311a573a4fc6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.193242] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93a37b1-5b47-4077-858a-4fa728daccb1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.235288] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b9f914-5302-430c-ab8a-f90e1b1aab2a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.244274] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "0041acd3-de76-4745-a1ab-c6563f774f33" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.244560] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "0041acd3-de76-4745-a1ab-c6563f774f33" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.244771] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "0041acd3-de76-4745-a1ab-c6563f774f33-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.244956] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "0041acd3-de76-4745-a1ab-c6563f774f33-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.245143] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "0041acd3-de76-4745-a1ab-c6563f774f33-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.247537] env[63293]: INFO nova.compute.manager [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Terminating instance [ 1074.249690] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "refresh_cache-0041acd3-de76-4745-a1ab-c6563f774f33" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.249981] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquired lock "refresh_cache-0041acd3-de76-4745-a1ab-c6563f774f33" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.250181] env[63293]: DEBUG nova.network.neutron [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1074.252916] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f936e4df-f016-49ea-93f9-9237a684e256 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.269354] env[63293]: DEBUG nova.compute.provider_tree [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.429123] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328236, 'name': Rename_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.634435] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52035aa5-04e5-e7c2-f347-ca019b0d03e1, 'name': SearchDatastore_Task, 'duration_secs': 0.032711} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.634858] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.634994] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 4969456a-be05-42d1-a4d5-44dc537f5a21/4969456a-be05-42d1-a4d5-44dc537f5a21.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1074.635379] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-635603ff-42a5-4e41-ad16-dc97b42741b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.644742] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for the task: (returnval){ [ 1074.644742] env[63293]: value = "task-1328237" [ 1074.644742] env[63293]: _type = "Task" [ 1074.644742] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.653108] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.772705] env[63293]: DEBUG nova.scheduler.client.report [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1074.777182] env[63293]: DEBUG nova.network.neutron [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1074.834464] env[63293]: DEBUG nova.network.neutron [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.929169] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328236, 'name': Rename_Task, 'duration_secs': 0.967099} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.929371] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1074.929651] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e9ed547-7140-40eb-912a-675430e73334 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.938413] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1074.938413] env[63293]: value = "task-1328238" [ 1074.938413] env[63293]: _type = "Task" [ 1074.938413] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.947459] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328238, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.075169] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1682d87e-1b14-4ba5-8418-8347cb0f4751 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.099458] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance 'bea1ea16-c224-4a0e-86ed-c2aaf920d3e1' progress to 0 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1075.157139] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328237, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.280542] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1075.280796] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.244s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.281107] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.337733] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Releasing lock "refresh_cache-0041acd3-de76-4745-a1ab-c6563f774f33" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.338134] env[63293]: DEBUG nova.compute.manager [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1075.338340] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1075.339319] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de926480-04dc-446d-92f0-86274ade9915 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.348274] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1075.348536] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-070557bb-b606-4749-adee-2c4e620587a4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.357193] env[63293]: DEBUG oslo_vmware.api [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1075.357193] env[63293]: value = "task-1328239" [ 1075.357193] env[63293]: _type = "Task" [ 1075.357193] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.366179] env[63293]: DEBUG oslo_vmware.api [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328239, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.449791] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328238, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.570817] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "97da107a-2e15-4da2-9197-7f85d960ff27" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.571078] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.605507] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1075.606094] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29376af9-976d-48e7-ae84-c67867c05657 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.615062] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1075.615062] env[63293]: value = "task-1328240" [ 1075.615062] env[63293]: _type = "Task" [ 1075.615062] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.624366] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.657756] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328237, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589253} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.658121] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] 4969456a-be05-42d1-a4d5-44dc537f5a21/4969456a-be05-42d1-a4d5-44dc537f5a21.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1075.658263] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1075.658503] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-58c47d26-64e3-4952-afae-65ec1f1d96a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.665842] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for the task: (returnval){ [ 1075.665842] env[63293]: value = "task-1328241" [ 1075.665842] env[63293]: _type = "Task" [ 1075.665842] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.673675] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328241, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.868206] env[63293]: DEBUG oslo_vmware.api [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328239, 'name': PowerOffVM_Task, 'duration_secs': 0.133795} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.868457] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1075.868633] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1075.869013] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1174300-a4d6-4878-839d-8e3dc66d2ff7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.913623] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1075.914206] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1075.914352] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Deleting the datastore file [datastore2] 0041acd3-de76-4745-a1ab-c6563f774f33 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1075.914533] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b8ce0ce-d25a-41b5-8591-f02fa910a7da {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.922925] env[63293]: DEBUG oslo_vmware.api [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for the task: (returnval){ [ 1075.922925] env[63293]: value = "task-1328243" [ 1075.922925] env[63293]: _type = "Task" [ 1075.922925] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.931944] env[63293]: DEBUG oslo_vmware.api [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328243, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.950188] env[63293]: DEBUG oslo_vmware.api [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328238, 'name': PowerOnVM_Task, 'duration_secs': 0.908065} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.950476] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1075.950688] env[63293]: INFO nova.compute.manager [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Took 8.41 seconds to spawn the instance on the hypervisor. [ 1075.950878] env[63293]: DEBUG nova.compute.manager [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1075.951735] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0c87b5-7070-40ba-9d68-ca89b34d68d5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.073928] env[63293]: DEBUG nova.compute.utils [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1076.126615] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328240, 'name': PowerOffVM_Task, 'duration_secs': 0.31059} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.126837] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1076.127033] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance 'bea1ea16-c224-4a0e-86ed-c2aaf920d3e1' progress to 17 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1076.177213] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328241, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108677} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.177543] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1076.178347] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d13d30-120d-4419-87f7-e9e0795494b5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.201627] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 4969456a-be05-42d1-a4d5-44dc537f5a21/4969456a-be05-42d1-a4d5-44dc537f5a21.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1076.201913] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce87f5d5-8d6b-449f-a661-d61b641e95dd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.223584] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for the task: (returnval){ [ 1076.223584] env[63293]: value = "task-1328244" [ 1076.223584] env[63293]: _type = "Task" [ 1076.223584] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.231795] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328244, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.433831] env[63293]: DEBUG oslo_vmware.api [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Task: {'id': task-1328243, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.343161} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.434095] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1076.434290] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1076.434475] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1076.434650] env[63293]: INFO nova.compute.manager [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1076.434885] env[63293]: DEBUG oslo.service.loopingcall [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.435088] env[63293]: DEBUG nova.compute.manager [-] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1076.435187] env[63293]: DEBUG nova.network.neutron [-] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1076.450795] env[63293]: DEBUG nova.network.neutron [-] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1076.473536] env[63293]: INFO nova.compute.manager [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Took 13.19 seconds to build instance. [ 1076.578792] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.633889] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:32:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1076.634175] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1076.634340] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1076.634564] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1076.634746] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1076.634903] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1076.635390] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1076.635800] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1076.636285] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1076.636567] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1076.636783] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1076.642032] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8ef746c-55aa-4599-8ec5-5c546c0c08aa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.660932] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1076.660932] env[63293]: value = "task-1328245" [ 1076.660932] env[63293]: _type = "Task" [ 1076.660932] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.670073] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328245, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.735369] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328244, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.783634] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.783830] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Starting heal instance info cache {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1076.783925] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Rebuilding the list of instances to heal {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1076.913055] env[63293]: INFO nova.compute.manager [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Rescuing [ 1076.913295] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.913461] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.914119] env[63293]: DEBUG nova.network.neutron [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1076.953160] env[63293]: DEBUG nova.network.neutron [-] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.975750] env[63293]: DEBUG oslo_concurrency.lockutils [None req-31aea644-6f26-455c-8da3-a4605a6203fa tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.695s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.172121] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328245, 'name': ReconfigVM_Task, 'duration_secs': 0.394474} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.172481] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance 'bea1ea16-c224-4a0e-86ed-c2aaf920d3e1' progress to 33 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1077.234703] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328244, 'name': ReconfigVM_Task, 'duration_secs': 0.862664} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.235147] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 4969456a-be05-42d1-a4d5-44dc537f5a21/4969456a-be05-42d1-a4d5-44dc537f5a21.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1077.235796] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a490fafe-a1e2-4f6d-9457-4ab438caca42 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.243680] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for the task: (returnval){ [ 1077.243680] env[63293]: value = "task-1328246" [ 1077.243680] env[63293]: _type = "Task" [ 1077.243680] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.252496] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328246, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.287952] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Skipping network cache update for instance because it is being deleted. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1077.288150] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Skipping network cache update for instance because it is Building. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1077.288395] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.288517] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquired lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.288648] env[63293]: DEBUG nova.network.neutron [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Forcefully refreshing network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1077.288792] env[63293]: DEBUG nova.objects.instance [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lazy-loading 'info_cache' on Instance uuid bea1ea16-c224-4a0e-86ed-c2aaf920d3e1 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.455762] env[63293]: INFO nova.compute.manager [-] [instance: 0041acd3-de76-4745-a1ab-c6563f774f33] Took 1.02 seconds to deallocate network for instance. [ 1077.640412] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "97da107a-2e15-4da2-9197-7f85d960ff27" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.640412] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.640652] env[63293]: INFO nova.compute.manager [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Attaching volume 86bf34c3-3f48-445f-9196-94e694cb61f4 to /dev/sdb [ 1077.649788] env[63293]: DEBUG nova.network.neutron [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Updating instance_info_cache with network_info: [{"id": "eef289b9-0d13-4ce5-be81-a5247187e457", "address": "fa:16:3e:e5:f5:fb", "network": {"id": "69a5064b-e3a0-4ef6-b722-09fbb9b50ca8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1786908568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b8adfd01c54641fb8bee8cd534d50569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef289b9-0d", "ovs_interfaceid": "eef289b9-0d13-4ce5-be81-a5247187e457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.673706] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cfe6cb0-f4c4-421e-b977-9518c58e8b4f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.678758] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1077.679050] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1077.679187] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1077.679381] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1077.679540] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1077.679692] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1077.679899] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1077.680077] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1077.680250] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1077.680415] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1077.680588] env[63293]: DEBUG nova.virt.hardware [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.685981] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfiguring VM instance instance-00000064 to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1077.686252] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d18f410c-80b0-49e3-90f3-e5a066346475 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.704861] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23aca02-3e35-48d0-bb49-c14ae66eefd0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.709160] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1077.709160] env[63293]: value = "task-1328247" [ 1077.709160] env[63293]: _type = "Task" [ 1077.709160] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.717786] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328247, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.724673] env[63293]: DEBUG nova.virt.block_device [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Updating existing volume attachment record: 76e45ff7-d64c-45f4-9f77-f0ed975ee6c4 {{(pid=63293) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1077.754553] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328246, 'name': Rename_Task, 'duration_secs': 0.427648} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.754850] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1077.755136] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-818e78ce-bcd0-44a7-965a-31a9abe23d49 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.764784] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for the task: (returnval){ [ 1077.764784] env[63293]: value = "task-1328248" [ 1077.764784] env[63293]: _type = "Task" [ 1077.764784] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.773904] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328248, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.967180] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.967513] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.967748] env[63293]: DEBUG nova.objects.instance [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lazy-loading 'resources' on Instance uuid 0041acd3-de76-4745-a1ab-c6563f774f33 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.152405] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Releasing lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.221706] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328247, 'name': ReconfigVM_Task, 'duration_secs': 0.418717} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.222029] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfigured VM instance instance-00000064 to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1078.222905] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae137ca-cb41-46a8-a219-10823657a418 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.248723] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] bea1ea16-c224-4a0e-86ed-c2aaf920d3e1/bea1ea16-c224-4a0e-86ed-c2aaf920d3e1.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1078.249123] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ab64b4d-6008-41ec-b018-f1df32e5aca4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.270395] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1078.270395] env[63293]: value = "task-1328250" [ 1078.270395] env[63293]: _type = "Task" [ 1078.270395] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.277132] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328248, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.282613] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328250, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.449044] env[63293]: DEBUG oslo_vmware.rw_handles [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525b9c46-0d4b-469c-eb2a-d9be683600bf/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1078.449845] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116a5382-2704-4c60-8c8c-325006b219dd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.456408] env[63293]: DEBUG oslo_vmware.rw_handles [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525b9c46-0d4b-469c-eb2a-d9be683600bf/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1078.456576] env[63293]: ERROR oslo_vmware.rw_handles [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525b9c46-0d4b-469c-eb2a-d9be683600bf/disk-0.vmdk due to incomplete transfer. [ 1078.456793] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c2872d30-9d80-4f43-a54c-3e3ed581b03c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.463947] env[63293]: DEBUG oslo_vmware.rw_handles [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525b9c46-0d4b-469c-eb2a-d9be683600bf/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1078.464161] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Uploaded image e4ccc7b7-9af6-4851-89da-fd23cc904ad3 to the Glance image server {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1078.466494] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Destroying the VM {{(pid=63293) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1078.466754] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4e6c887f-965d-451e-950c-c9ca8690056d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.473990] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1078.473990] env[63293]: value = "task-1328251" [ 1078.473990] env[63293]: _type = "Task" [ 1078.473990] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.482409] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328251, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.687023] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1078.687283] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95de6ac3-4f8d-4cb4-a490-6e3a36655886 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.699403] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1078.699403] env[63293]: value = "task-1328252" [ 1078.699403] env[63293]: _type = "Task" [ 1078.699403] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.708506] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328252, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.710283] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6d8d97-0380-4361-a93e-3f34c13bc7f0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.718639] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad897ec-f5f0-4f71-957b-3a43951168e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.752632] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72eca7a-31c9-4294-abdc-30e4327b9399 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.760897] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0130be-5b98-4354-8f7e-9ae6e8b83431 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.777173] env[63293]: DEBUG nova.compute.provider_tree [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.785125] env[63293]: DEBUG oslo_vmware.api [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328248, 'name': PowerOnVM_Task, 'duration_secs': 0.994403} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.786328] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1078.786610] env[63293]: INFO nova.compute.manager [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Took 9.01 seconds to spawn the instance on the hypervisor. [ 1078.786817] env[63293]: DEBUG nova.compute.manager [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1078.787902] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2be35e8-35aa-470d-ab89-a81d10601d3a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.793734] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328250, 'name': ReconfigVM_Task, 'duration_secs': 0.451469} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.794317] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfigured VM instance instance-00000064 to attach disk [datastore2] bea1ea16-c224-4a0e-86ed-c2aaf920d3e1/bea1ea16-c224-4a0e-86ed-c2aaf920d3e1.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.794631] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance 'bea1ea16-c224-4a0e-86ed-c2aaf920d3e1' progress to 50 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1078.983575] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328251, 'name': Destroy_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.003755] env[63293]: DEBUG nova.network.neutron [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance_info_cache with network_info: [{"id": "c589de77-25fc-43f7-9480-e6c97bbda93d", "address": "fa:16:3e:4a:ef:d2", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc589de77-25", "ovs_interfaceid": "c589de77-25fc-43f7-9480-e6c97bbda93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.209565] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328252, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.286399] env[63293]: DEBUG nova.scheduler.client.report [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.309920] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91481b5d-4c91-4c29-8042-21785591369e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.312585] env[63293]: INFO nova.compute.manager [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Took 14.80 seconds to build instance. [ 1079.335027] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aef5a07-eb7a-40a7-8d98-f184cb63c4e1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.356271] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance 'bea1ea16-c224-4a0e-86ed-c2aaf920d3e1' progress to 67 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1079.484562] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328251, 'name': Destroy_Task, 'duration_secs': 0.771851} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.484890] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Destroyed the VM [ 1079.485150] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Deleting Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1079.485416] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c708319e-c233-4b1e-bee4-20aca9708e19 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.492516] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1079.492516] env[63293]: value = "task-1328253" [ 1079.492516] env[63293]: _type = "Task" [ 1079.492516] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.500128] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328253, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.505684] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Releasing lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.505876] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updated the network info_cache for instance {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1079.506071] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.506231] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.506376] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.506525] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.506685] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.506809] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Cleaning up deleted instances {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1079.711024] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328252, 'name': PowerOffVM_Task, 'duration_secs': 0.924493} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.711535] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1079.712327] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e44a77-8ed8-4c65-9efa-4fc047f2c2ac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.736352] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74af7f85-5829-479c-bd41-df1b7ca9156c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.773678] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1079.774051] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-712b9c2c-933f-4bd6-9603-22cb760f93e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.785199] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1079.785199] env[63293]: value = "task-1328254" [ 1079.785199] env[63293]: _type = "Task" [ 1079.785199] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.794259] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.796140] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328254, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.815554] env[63293]: DEBUG oslo_concurrency.lockutils [None req-96240dde-da55-4bc1-b94c-36386479745e tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Lock "4969456a-be05-42d1-a4d5-44dc537f5a21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.310s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.821499] env[63293]: INFO nova.scheduler.client.report [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Deleted allocations for instance 0041acd3-de76-4745-a1ab-c6563f774f33 [ 1080.004833] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328253, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.020264] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] There are 48 instances to clean {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1080.020606] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: b0171e46-ec0b-4fd6-bb04-6f943757b083] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.296975] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] VM already powered off {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1080.297420] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1080.297826] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.298135] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.298462] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1080.298767] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dcd44979-e2e6-43fd-b378-a8868e10ec73 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.308018] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1080.308214] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1080.308976] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e47e05b3-0bad-4783-85d3-8bc355593314 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.314460] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1080.314460] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52aa622a-b41b-6a64-2740-3882fcaed982" [ 1080.314460] env[63293]: _type = "Task" [ 1080.314460] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.322803] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52aa622a-b41b-6a64-2740-3882fcaed982, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.328752] env[63293]: DEBUG oslo_concurrency.lockutils [None req-65ca67f6-c9dd-49eb-b709-6a9e27dd2f7e tempest-ServerShowV247Test-1491806081 tempest-ServerShowV247Test-1491806081-project-member] Lock "0041acd3-de76-4745-a1ab-c6563f774f33" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.084s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.380597] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquiring lock "4969456a-be05-42d1-a4d5-44dc537f5a21" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.380872] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Lock "4969456a-be05-42d1-a4d5-44dc537f5a21" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.381109] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquiring lock "4969456a-be05-42d1-a4d5-44dc537f5a21-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.381350] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Lock "4969456a-be05-42d1-a4d5-44dc537f5a21-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.381557] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Lock "4969456a-be05-42d1-a4d5-44dc537f5a21-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.383744] env[63293]: INFO nova.compute.manager [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Terminating instance [ 1080.385486] env[63293]: DEBUG nova.compute.manager [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1080.385685] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1080.386517] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652e74f7-13a1-4f59-af3e-c29625d2ae62 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.394784] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.395046] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-228e69c6-814e-495c-bd60-049515240f37 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.401650] env[63293]: DEBUG oslo_vmware.api [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for the task: (returnval){ [ 1080.401650] env[63293]: value = "task-1328256" [ 1080.401650] env[63293]: _type = "Task" [ 1080.401650] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.412506] env[63293]: DEBUG oslo_vmware.api [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328256, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.503922] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328253, 'name': RemoveSnapshot_Task, 'duration_secs': 0.694046} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.504225] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Deleted Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1080.504512] env[63293]: DEBUG nova.compute.manager [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1080.505602] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b55e53ff-5322-4c85-bf18-9a4f15bca35d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.523805] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 46a9329b-6e3b-454f-8a50-30ff7bc719f0] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.825980] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52aa622a-b41b-6a64-2740-3882fcaed982, 'name': SearchDatastore_Task, 'duration_secs': 0.009631} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.826815] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0164857a-d941-4c0b-878d-5da3dc95dd76 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.833033] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1080.833033] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d07654-fe79-9dc9-ec6d-1062c2aa3770" [ 1080.833033] env[63293]: _type = "Task" [ 1080.833033] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.841543] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d07654-fe79-9dc9-ec6d-1062c2aa3770, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.913073] env[63293]: DEBUG oslo_vmware.api [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328256, 'name': PowerOffVM_Task, 'duration_secs': 0.215209} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.913387] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1080.913562] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1080.913851] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef98a2a5-abae-4338-8417-f6c1f4690896 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.983146] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1080.983334] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1080.983529] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Deleting the datastore file [datastore1] 4969456a-be05-42d1-a4d5-44dc537f5a21 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1080.983827] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-297e339b-3b25-4160-91c4-f3ce5d18a308 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.990985] env[63293]: DEBUG oslo_vmware.api [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for the task: (returnval){ [ 1080.990985] env[63293]: value = "task-1328258" [ 1080.990985] env[63293]: _type = "Task" [ 1080.990985] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.000442] env[63293]: DEBUG oslo_vmware.api [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328258, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.002571] env[63293]: DEBUG nova.network.neutron [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Port c589de77-25fc-43f7-9480-e6c97bbda93d binding to destination host cpu-1 is already ACTIVE {{(pid=63293) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1081.020085] env[63293]: INFO nova.compute.manager [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Shelve offloading [ 1081.022077] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1081.023211] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc4dcc44-b4dc-4ed7-baed-0a57d0ef028b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.026540] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 2bf4a8a9-c46d-4512-96b7-b34b1bf69f50] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.030771] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1081.030771] env[63293]: value = "task-1328259" [ 1081.030771] env[63293]: _type = "Task" [ 1081.030771] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.043246] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] VM already powered off {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1081.043457] env[63293]: DEBUG nova.compute.manager [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1081.044267] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9690c704-411c-4178-a8d1-392f531b1122 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.051958] env[63293]: DEBUG oslo_concurrency.lockutils [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.051958] env[63293]: DEBUG oslo_concurrency.lockutils [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.051958] env[63293]: DEBUG nova.network.neutron [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1081.343332] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d07654-fe79-9dc9-ec6d-1062c2aa3770, 'name': SearchDatastore_Task, 'duration_secs': 0.011479} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.343597] env[63293]: DEBUG oslo_concurrency.lockutils [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.343861] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] ba8404c7-b43e-4776-941e-4d9bd8fb4c9d/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk. {{(pid=63293) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1081.344139] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e1b7944-4573-4aea-9657-769b7af4e277 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.351092] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1081.351092] env[63293]: value = "task-1328260" [ 1081.351092] env[63293]: _type = "Task" [ 1081.351092] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.359332] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328260, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.501360] env[63293]: DEBUG oslo_vmware.api [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Task: {'id': task-1328258, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.27378} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.501683] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.501925] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1081.502173] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1081.502448] env[63293]: INFO nova.compute.manager [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1081.502740] env[63293]: DEBUG oslo.service.loopingcall [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1081.502981] env[63293]: DEBUG nova.compute.manager [-] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1081.503122] env[63293]: DEBUG nova.network.neutron [-] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1081.529691] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 71154ca4-ea2c-4187-8319-0bfb9b289eab] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.866697] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328260, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.875038] env[63293]: DEBUG nova.compute.manager [req-32720af2-13bf-4357-ba7e-e74f8245bc7a req-9e505fd1-7920-458a-a55b-4b722a009af4 service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Received event network-vif-deleted-58d85ce7-565c-471e-abde-4f572863d75a {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.875439] env[63293]: INFO nova.compute.manager [req-32720af2-13bf-4357-ba7e-e74f8245bc7a req-9e505fd1-7920-458a-a55b-4b722a009af4 service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Neutron deleted interface 58d85ce7-565c-471e-abde-4f572863d75a; detaching it from the instance and deleting it from the info cache [ 1081.875798] env[63293]: DEBUG nova.network.neutron [req-32720af2-13bf-4357-ba7e-e74f8245bc7a req-9e505fd1-7920-458a-a55b-4b722a009af4 service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.899945] env[63293]: DEBUG nova.network.neutron [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updating instance_info_cache with network_info: [{"id": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "address": "fa:16:3e:85:83:c3", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3313c3a6-17", "ovs_interfaceid": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.026411] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.026411] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.026678] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.035176] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: bdcc5cb2-0ebd-4b55-8c48-47f927a11ea7] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.273227] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Volume attach. Driver type: vmdk {{(pid=63293) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1082.273482] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283852', 'volume_id': '86bf34c3-3f48-445f-9196-94e694cb61f4', 'name': 'volume-86bf34c3-3f48-445f-9196-94e694cb61f4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '97da107a-2e15-4da2-9197-7f85d960ff27', 'attached_at': '', 'detached_at': '', 'volume_id': '86bf34c3-3f48-445f-9196-94e694cb61f4', 'serial': '86bf34c3-3f48-445f-9196-94e694cb61f4'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1082.274429] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5c3981-a6c3-400f-9052-384d59952fb9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.292890] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f8f1e3-a701-46c9-b8ae-8e2f6754f60d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.316581] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] volume-86bf34c3-3f48-445f-9196-94e694cb61f4/volume-86bf34c3-3f48-445f-9196-94e694cb61f4.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.316819] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eede9b70-ef8d-40a2-a1b4-51487f92cca3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.334553] env[63293]: DEBUG oslo_vmware.api [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1082.334553] env[63293]: value = "task-1328261" [ 1082.334553] env[63293]: _type = "Task" [ 1082.334553] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.342015] env[63293]: DEBUG oslo_vmware.api [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328261, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.344355] env[63293]: DEBUG nova.network.neutron [-] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.364815] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328260, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.764923} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.365368] env[63293]: INFO nova.virt.vmwareapi.ds_util [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore1] ba8404c7-b43e-4776-941e-4d9bd8fb4c9d/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk. [ 1082.366173] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a7aafc-6374-44a1-bcbb-1e513b098c46 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.392450] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] ba8404c7-b43e-4776-941e-4d9bd8fb4c9d/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.392773] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed32f317-f014-4805-882a-2560ab5344f8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.394754] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7f2ae8b-51a2-4f95-9199-91426d132efd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.408196] env[63293]: DEBUG oslo_concurrency.lockutils [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.418250] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1082.418250] env[63293]: value = "task-1328262" [ 1082.418250] env[63293]: _type = "Task" [ 1082.418250] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.425287] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75deffa-aacf-4df6-b3c2-8e8743fdfc52 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.449935] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328262, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.456903] env[63293]: DEBUG nova.compute.manager [req-32720af2-13bf-4357-ba7e-e74f8245bc7a req-9e505fd1-7920-458a-a55b-4b722a009af4 service nova] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Detach interface failed, port_id=58d85ce7-565c-471e-abde-4f572863d75a, reason: Instance 4969456a-be05-42d1-a4d5-44dc537f5a21 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1082.538037] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: d605e066-806a-4aac-9af6-f79d542ab2e3] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.755405] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1082.756469] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872e5c8d-5950-44bb-b0f2-a3faab9578c1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.765011] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1082.765314] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-220e5bc7-7269-44a3-b928-eb86fb3a090d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.836149] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1082.836353] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1082.836587] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleting the datastore file [datastore2] 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1082.839828] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd1784ff-dd94-4809-831f-9b28cbe76469 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.847067] env[63293]: INFO nova.compute.manager [-] [instance: 4969456a-be05-42d1-a4d5-44dc537f5a21] Took 1.34 seconds to deallocate network for instance. [ 1082.847379] env[63293]: DEBUG oslo_vmware.api [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328261, 'name': ReconfigVM_Task, 'duration_secs': 0.377009} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.850060] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Reconfigured VM instance instance-00000066 to attach disk [datastore1] volume-86bf34c3-3f48-445f-9196-94e694cb61f4/volume-86bf34c3-3f48-445f-9196-94e694cb61f4.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.854951] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1082.854951] env[63293]: value = "task-1328264" [ 1082.854951] env[63293]: _type = "Task" [ 1082.854951] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.857711] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b287daf0-59d0-469d-b1ab-79300bbba55d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.876084] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328264, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.877389] env[63293]: DEBUG oslo_vmware.api [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1082.877389] env[63293]: value = "task-1328265" [ 1082.877389] env[63293]: _type = "Task" [ 1082.877389] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.885075] env[63293]: DEBUG oslo_vmware.api [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328265, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.929703] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328262, 'name': ReconfigVM_Task, 'duration_secs': 0.35221} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.930021] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Reconfigured VM instance instance-0000006b to attach disk [datastore1] ba8404c7-b43e-4776-941e-4d9bd8fb4c9d/1427aa66-93c7-49c6-9e28-dc7fe851dced-rescue.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.930907] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba11ae72-95e6-4766-841e-0b262b7a458f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.956890] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b137a9f-428e-485a-9bb0-f6c4c41bbf12 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.973098] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1082.973098] env[63293]: value = "task-1328266" [ 1082.973098] env[63293]: _type = "Task" [ 1082.973098] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.981240] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328266, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.041243] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 9937a3c4-c3a6-423b-ac4f-50b7ec5879d2] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.064531] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.064732] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.064912] env[63293]: DEBUG nova.network.neutron [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1083.369574] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.369832] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.370053] env[63293]: DEBUG nova.objects.instance [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Lazy-loading 'resources' on Instance uuid 4969456a-be05-42d1-a4d5-44dc537f5a21 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.380507] env[63293]: DEBUG oslo_vmware.api [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328264, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19514} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.383405] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1083.383597] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1083.383775] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1083.390898] env[63293]: DEBUG oslo_vmware.api [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328265, 'name': ReconfigVM_Task, 'duration_secs': 0.148651} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.391187] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283852', 'volume_id': '86bf34c3-3f48-445f-9196-94e694cb61f4', 'name': 'volume-86bf34c3-3f48-445f-9196-94e694cb61f4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '97da107a-2e15-4da2-9197-7f85d960ff27', 'attached_at': '', 'detached_at': '', 'volume_id': '86bf34c3-3f48-445f-9196-94e694cb61f4', 'serial': '86bf34c3-3f48-445f-9196-94e694cb61f4'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1083.413917] env[63293]: INFO nova.scheduler.client.report [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleted allocations for instance 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d [ 1083.483309] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328266, 'name': ReconfigVM_Task, 'duration_secs': 0.229909} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.483570] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1083.483838] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80b4ce02-2575-43fd-af6d-37db6884dee4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.490569] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1083.490569] env[63293]: value = "task-1328267" [ 1083.490569] env[63293]: _type = "Task" [ 1083.490569] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.499663] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328267, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.544731] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 9d3fcf27-2c2c-4975-a802-a5c0859640d0] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.781769] env[63293]: DEBUG nova.network.neutron [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance_info_cache with network_info: [{"id": "c589de77-25fc-43f7-9480-e6c97bbda93d", "address": "fa:16:3e:4a:ef:d2", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc589de77-25", "ovs_interfaceid": "c589de77-25fc-43f7-9480-e6c97bbda93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.905216] env[63293]: DEBUG nova.compute.manager [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Received event network-vif-unplugged-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.905498] env[63293]: DEBUG oslo_concurrency.lockutils [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] Acquiring lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.905666] env[63293]: DEBUG oslo_concurrency.lockutils [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.905859] env[63293]: DEBUG oslo_concurrency.lockutils [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.905990] env[63293]: DEBUG nova.compute.manager [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] No waiting events found dispatching network-vif-unplugged-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1083.906203] env[63293]: WARNING nova.compute.manager [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Received unexpected event network-vif-unplugged-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 for instance with vm_state shelved_offloaded and task_state None. [ 1083.906462] env[63293]: DEBUG nova.compute.manager [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Received event network-changed-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.906582] env[63293]: DEBUG nova.compute.manager [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Refreshing instance network info cache due to event network-changed-3313c3a6-17f4-4c97-b19f-9edfd6eb4099. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1083.906774] env[63293]: DEBUG oslo_concurrency.lockutils [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] Acquiring lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.906915] env[63293]: DEBUG oslo_concurrency.lockutils [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] Acquired lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.907087] env[63293]: DEBUG nova.network.neutron [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Refreshing network info cache for port 3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1083.917781] env[63293]: DEBUG oslo_concurrency.lockutils [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.964029] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a6f79a-854b-4259-b229-a9df92a8270a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.972119] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d897dcf-d784-4b90-9ad2-c58e316d5c6b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.005244] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-040d55dd-6827-4a51-8e99-9882b75ebf66 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.015145] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305b9440-72fe-4edb-a003-8e3cbed71384 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.018655] env[63293]: DEBUG oslo_vmware.api [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328267, 'name': PowerOnVM_Task, 'duration_secs': 0.437831} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.018914] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1084.021711] env[63293]: DEBUG nova.compute.manager [None req-a55db723-20cd-4742-8c47-be0b1613daf5 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1084.022487] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a611ea-5f44-4961-a254-e8c8df9cdae9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.033842] env[63293]: DEBUG nova.compute.provider_tree [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.047361] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 657d5f21-b985-4e3c-8799-b3aae2b1c509] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.284690] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.430096] env[63293]: DEBUG nova.objects.instance [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lazy-loading 'flavor' on Instance uuid 97da107a-2e15-4da2-9197-7f85d960ff27 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.537638] env[63293]: DEBUG nova.scheduler.client.report [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.549999] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 66be1abd-c541-4b8a-8d72-98bde03b1888] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.715195] env[63293]: DEBUG nova.network.neutron [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updated VIF entry in instance network info cache for port 3313c3a6-17f4-4c97-b19f-9edfd6eb4099. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1084.715581] env[63293]: DEBUG nova.network.neutron [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updating instance_info_cache with network_info: [{"id": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "address": "fa:16:3e:85:83:c3", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap3313c3a6-17", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.793929] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa373d66-757b-4826-b07e-84495a614542 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.801868] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6e3692-0135-4f8e-99b4-51b05f8036b0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.885508] env[63293]: INFO nova.compute.manager [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Unrescuing [ 1084.885801] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.885965] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquired lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.886159] env[63293]: DEBUG nova.network.neutron [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1084.934865] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5811dc5-dcca-40c1-98a9-eaa2394c1cf5 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.294s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.045278] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.675s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.048620] env[63293]: DEBUG oslo_concurrency.lockutils [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.130s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.048895] env[63293]: DEBUG nova.objects.instance [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lazy-loading 'resources' on Instance uuid 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1085.052992] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: f9ccb216-1435-44c5-ab34-a6388d794551] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.066880] env[63293]: INFO nova.scheduler.client.report [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Deleted allocations for instance 4969456a-be05-42d1-a4d5-44dc537f5a21 [ 1085.218190] env[63293]: DEBUG oslo_concurrency.lockutils [req-6c44b5da-e0b0-4b1b-906a-8cbc82586474 req-236a2830-cc3e-453d-af58-0334cb34e5ad service nova] Releasing lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.338547] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "97da107a-2e15-4da2-9197-7f85d960ff27" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.338825] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.551579] env[63293]: DEBUG nova.objects.instance [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lazy-loading 'numa_topology' on Instance uuid 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1085.555302] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: f85a1fd2-db4f-4c6b-8fc8-47c3cf9afc8a] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.573914] env[63293]: DEBUG oslo_concurrency.lockutils [None req-f4459025-3c94-4455-a4bb-aac050e00df7 tempest-ServersNegativeTestMultiTenantJSON-445112376 tempest-ServersNegativeTestMultiTenantJSON-445112376-project-member] Lock "4969456a-be05-42d1-a4d5-44dc537f5a21" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.193s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.751552] env[63293]: DEBUG nova.network.neutron [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Updating instance_info_cache with network_info: [{"id": "eef289b9-0d13-4ce5-be81-a5247187e457", "address": "fa:16:3e:e5:f5:fb", "network": {"id": "69a5064b-e3a0-4ef6-b722-09fbb9b50ca8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1786908568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b8adfd01c54641fb8bee8cd534d50569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef289b9-0d", "ovs_interfaceid": "eef289b9-0d13-4ce5-be81-a5247187e457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.842073] env[63293]: INFO nova.compute.manager [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Detaching volume 86bf34c3-3f48-445f-9196-94e694cb61f4 [ 1085.880282] env[63293]: INFO nova.virt.block_device [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Attempting to driver detach volume 86bf34c3-3f48-445f-9196-94e694cb61f4 from mountpoint /dev/sdb [ 1085.880282] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Volume detach. Driver type: vmdk {{(pid=63293) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1085.880282] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283852', 'volume_id': '86bf34c3-3f48-445f-9196-94e694cb61f4', 'name': 'volume-86bf34c3-3f48-445f-9196-94e694cb61f4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '97da107a-2e15-4da2-9197-7f85d960ff27', 'attached_at': '', 'detached_at': '', 'volume_id': '86bf34c3-3f48-445f-9196-94e694cb61f4', 'serial': '86bf34c3-3f48-445f-9196-94e694cb61f4'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1085.880282] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3381ed2-6ea1-48c2-83d3-21acb535f277 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.905840] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2050ac34-97c7-40d4-8360-eb818dfc35db {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.914889] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6e3c10-5c32-4d47-9511-8989bb96ad1e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.937678] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf1dd73-3207-495c-8251-8c5314db492d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.941032] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11d70d5-6e7c-46e2-abaa-b97daf6c0e0c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.959707] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] The volume has not been displaced from its original location: [datastore1] volume-86bf34c3-3f48-445f-9196-94e694cb61f4/volume-86bf34c3-3f48-445f-9196-94e694cb61f4.vmdk. No consolidation needed. {{(pid=63293) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1085.965595] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1085.983616] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1638c311-0168-4897-9d96-69f46bf3e88f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.999221] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00401984-835a-4941-bff4-cdfb27a8e286 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.008805] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance 'bea1ea16-c224-4a0e-86ed-c2aaf920d3e1' progress to 83 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1086.014286] env[63293]: DEBUG oslo_vmware.api [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1086.014286] env[63293]: value = "task-1328268" [ 1086.014286] env[63293]: _type = "Task" [ 1086.014286] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.024639] env[63293]: DEBUG oslo_vmware.api [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328268, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.054528] env[63293]: DEBUG nova.objects.base [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Object Instance<97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d> lazy-loaded attributes: resources,numa_topology {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1086.057541] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 6f1fb925-72c2-4309-b3fd-aea217d38bf2] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.142450] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca74706c-a31e-4cc7-a45b-d19784666fcc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.151404] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c0cb7d-0b78-4f15-84fe-c173460944c8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.184345] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfca117a-bb2b-410c-a2b6-ed0178821954 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.193346] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057d3f3a-3561-45ed-8b4c-f86d002fb897 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.211050] env[63293]: DEBUG nova.compute.provider_tree [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.254673] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Releasing lock "refresh_cache-ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.255371] env[63293]: DEBUG nova.objects.instance [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lazy-loading 'flavor' on Instance uuid ba8404c7-b43e-4776-941e-4d9bd8fb4c9d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.362966] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.517801] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1086.518202] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d25b295-6f3b-4c32-ade5-6929b4db9dfd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.530127] env[63293]: DEBUG oslo_vmware.api [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328268, 'name': ReconfigVM_Task, 'duration_secs': 0.274036} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.531603] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1086.537655] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1086.537655] env[63293]: value = "task-1328269" [ 1086.537655] env[63293]: _type = "Task" [ 1086.537655] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.537992] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36ccba46-c60d-42d4-9cb8-3c06b521a5c0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.559299] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328269, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.560798] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 828180c0-b7f8-4666-9f5d-e6741ef6495a] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.562725] env[63293]: DEBUG oslo_vmware.api [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1086.562725] env[63293]: value = "task-1328270" [ 1086.562725] env[63293]: _type = "Task" [ 1086.562725] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.571743] env[63293]: DEBUG oslo_vmware.api [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328270, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.714673] env[63293]: DEBUG nova.scheduler.client.report [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1086.761821] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1bbbd5-bd90-4a3d-9b70-6b08231504c6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.786170] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1086.786551] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9177335-b60e-4d7b-8e85-490eec2d7600 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.795042] env[63293]: DEBUG oslo_vmware.api [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1086.795042] env[63293]: value = "task-1328271" [ 1086.795042] env[63293]: _type = "Task" [ 1086.795042] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.806069] env[63293]: DEBUG oslo_vmware.api [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.059076] env[63293]: DEBUG oslo_vmware.api [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328269, 'name': PowerOnVM_Task, 'duration_secs': 0.469491} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.059499] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1087.059499] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d384a34c-c73e-4aaa-8ac8-021f1fc86e0d tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance 'bea1ea16-c224-4a0e-86ed-c2aaf920d3e1' progress to 100 {{(pid=63293) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1087.063374] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 2f55ac2a-63d1-4713-ab0c-47fb6cf7d535] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.090557] env[63293]: DEBUG oslo_vmware.api [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328270, 'name': ReconfigVM_Task, 'duration_secs': 0.165098} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.090750] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283852', 'volume_id': '86bf34c3-3f48-445f-9196-94e694cb61f4', 'name': 'volume-86bf34c3-3f48-445f-9196-94e694cb61f4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '97da107a-2e15-4da2-9197-7f85d960ff27', 'attached_at': '', 'detached_at': '', 'volume_id': '86bf34c3-3f48-445f-9196-94e694cb61f4', 'serial': '86bf34c3-3f48-445f-9196-94e694cb61f4'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1087.219292] env[63293]: DEBUG oslo_concurrency.lockutils [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.171s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.305146] env[63293]: DEBUG oslo_vmware.api [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328271, 'name': PowerOffVM_Task, 'duration_secs': 0.264447} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.305480] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1087.310747] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1087.310991] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e8d8809-1c21-4680-84ac-e59213d78c2d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.331225] env[63293]: DEBUG oslo_vmware.api [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1087.331225] env[63293]: value = "task-1328272" [ 1087.331225] env[63293]: _type = "Task" [ 1087.331225] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.339213] env[63293]: DEBUG oslo_vmware.api [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328272, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.572198] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 35cbf277-b85a-4101-beae-0125fb5a61f5] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.636199] env[63293]: DEBUG nova.objects.instance [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lazy-loading 'flavor' on Instance uuid 97da107a-2e15-4da2-9197-7f85d960ff27 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.731281] env[63293]: DEBUG oslo_concurrency.lockutils [None req-836f683c-89f7-44f4-9328-9cfbdbb434dd tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.600s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.731798] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.369s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.731986] env[63293]: INFO nova.compute.manager [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Unshelving [ 1087.842275] env[63293]: DEBUG oslo_vmware.api [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328272, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.074252] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: f429869f-9e9f-443e-a980-67a80c3bd799] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.346510] env[63293]: DEBUG oslo_vmware.api [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328272, 'name': ReconfigVM_Task, 'duration_secs': 0.578947} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.346510] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1088.346510] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1088.346696] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1ed683a-4211-4c50-831f-f2c39ebd6e84 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.354472] env[63293]: DEBUG oslo_vmware.api [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1088.354472] env[63293]: value = "task-1328273" [ 1088.354472] env[63293]: _type = "Task" [ 1088.354472] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.362145] env[63293]: DEBUG oslo_vmware.api [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328273, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.578332] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 2557bded-6ccb-4c0f-8b97-34529b4bb740] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.642794] env[63293]: DEBUG oslo_concurrency.lockutils [None req-cbf2c856-5f06-4b7f-a058-f2700cd55992 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.304s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.762671] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.763057] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.763301] env[63293]: DEBUG nova.objects.instance [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lazy-loading 'pci_requests' on Instance uuid 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.864784] env[63293]: DEBUG oslo_vmware.api [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328273, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.081912] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: fe55a9b7-28b5-4b04-84f1-6b85b579a144] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.267831] env[63293]: DEBUG nova.objects.instance [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lazy-loading 'numa_topology' on Instance uuid 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.364853] env[63293]: DEBUG oslo_vmware.api [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328273, 'name': PowerOnVM_Task, 'duration_secs': 0.567635} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.365108] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1089.365346] env[63293]: DEBUG nova.compute.manager [None req-c4ba49d1-ac04-4180-b41b-c2c37d1544b7 tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1089.366140] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03786cf2-fe35-4440-961b-e94a8c9e4bb0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.584692] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 8625a358-ec48-46b9-8738-c49a2ba58362] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.734387] env[63293]: DEBUG oslo_concurrency.lockutils [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "97da107a-2e15-4da2-9197-7f85d960ff27" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.734623] env[63293]: DEBUG oslo_concurrency.lockutils [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.734868] env[63293]: DEBUG oslo_concurrency.lockutils [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "97da107a-2e15-4da2-9197-7f85d960ff27-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.735071] env[63293]: DEBUG oslo_concurrency.lockutils [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.735268] env[63293]: DEBUG oslo_concurrency.lockutils [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.737964] env[63293]: INFO nova.compute.manager [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Terminating instance [ 1089.739772] env[63293]: DEBUG nova.compute.manager [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1089.740041] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1089.740983] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271018d0-1b2e-4cf1-b670-64898495931c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.749274] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1089.749529] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4c66489-97db-4ed0-b65e-96c80cbe9596 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.756176] env[63293]: DEBUG oslo_vmware.api [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1089.756176] env[63293]: value = "task-1328274" [ 1089.756176] env[63293]: _type = "Task" [ 1089.756176] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.765115] env[63293]: DEBUG oslo_vmware.api [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328274, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.771912] env[63293]: INFO nova.compute.claims [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1089.939873] env[63293]: DEBUG nova.network.neutron [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Port c589de77-25fc-43f7-9480-e6c97bbda93d binding to destination host cpu-1 is already ACTIVE {{(pid=63293) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1089.940179] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.940339] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.940516] env[63293]: DEBUG nova.network.neutron [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1090.088601] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 5b1264b4-2554-4c62-9ff9-4616b03d3609] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.267525] env[63293]: DEBUG oslo_vmware.api [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328274, 'name': PowerOffVM_Task, 'duration_secs': 0.254644} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.267804] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1090.267979] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1090.268254] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-58b2ad1c-52d8-4c6a-aa4e-0f31c196d08e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.351338] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1090.352730] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1090.352730] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleting the datastore file [datastore1] 97da107a-2e15-4da2-9197-7f85d960ff27 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1090.353092] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-06dd13f4-8994-4dea-a2b7-2c1eb07bcd94 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.361218] env[63293]: DEBUG oslo_vmware.api [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1090.361218] env[63293]: value = "task-1328276" [ 1090.361218] env[63293]: _type = "Task" [ 1090.361218] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.369219] env[63293]: DEBUG oslo_vmware.api [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328276, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.591557] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 5a412c9d-35d0-4fcf-88b3-10a6bcfdef5d] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.644183] env[63293]: DEBUG nova.network.neutron [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance_info_cache with network_info: [{"id": "c589de77-25fc-43f7-9480-e6c97bbda93d", "address": "fa:16:3e:4a:ef:d2", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc589de77-25", "ovs_interfaceid": "c589de77-25fc-43f7-9480-e6c97bbda93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.843372] env[63293]: DEBUG oslo_concurrency.lockutils [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.843607] env[63293]: DEBUG oslo_concurrency.lockutils [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.843810] env[63293]: DEBUG oslo_concurrency.lockutils [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.843997] env[63293]: DEBUG oslo_concurrency.lockutils [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.844183] env[63293]: DEBUG oslo_concurrency.lockutils [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.846323] env[63293]: INFO nova.compute.manager [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Terminating instance [ 1090.848311] env[63293]: DEBUG nova.compute.manager [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1090.848508] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1090.849363] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f141f359-619f-4810-bfc0-c5daf1c0a640 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.860110] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1090.860110] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eabafe4c-5e2e-4228-9f8e-9b71a0d74e3c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.871244] env[63293]: DEBUG oslo_vmware.api [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1090.871244] env[63293]: value = "task-1328277" [ 1090.871244] env[63293]: _type = "Task" [ 1090.871244] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.874736] env[63293]: DEBUG oslo_vmware.api [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328276, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.475012} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.877653] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1090.877851] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1090.878056] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1090.878239] env[63293]: INFO nova.compute.manager [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1090.878482] env[63293]: DEBUG oslo.service.loopingcall [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1090.879532] env[63293]: DEBUG nova.compute.manager [-] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1090.879636] env[63293]: DEBUG nova.network.neutron [-] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1090.881850] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aacaffa1-0401-458c-91cd-31c94c0bcc8d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.894094] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe501a9-58bc-4b62-b7c7-63b8c385648b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.897141] env[63293]: DEBUG oslo_vmware.api [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.927150] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd0bf86-c447-4aa8-b086-3b025d1a9140 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.936425] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a554cb-ef88-41ca-91bb-8e30a4baa855 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.952341] env[63293]: DEBUG nova.compute.provider_tree [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.094534] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: eec9bb88-c5c4-4d8a-8a30-84b8d89e153d] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.147029] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.382150] env[63293]: DEBUG nova.compute.manager [req-e923c9f9-641a-4599-9f4e-e8883af414b6 req-9d891425-27c6-462c-8974-c385d8b9e23f service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Received event network-vif-deleted-a6a90b56-3edd-4270-a9f8-181df8dbb637 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1091.382346] env[63293]: INFO nova.compute.manager [req-e923c9f9-641a-4599-9f4e-e8883af414b6 req-9d891425-27c6-462c-8974-c385d8b9e23f service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Neutron deleted interface a6a90b56-3edd-4270-a9f8-181df8dbb637; detaching it from the instance and deleting it from the info cache [ 1091.382514] env[63293]: DEBUG nova.network.neutron [req-e923c9f9-641a-4599-9f4e-e8883af414b6 req-9d891425-27c6-462c-8974-c385d8b9e23f service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.387412] env[63293]: DEBUG oslo_vmware.api [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328277, 'name': PowerOffVM_Task, 'duration_secs': 0.296836} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.388028] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1091.388028] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1091.388146] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f6e34bb9-4ca5-43f9-9e4c-e61eb23bdc7d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.458155] env[63293]: DEBUG nova.scheduler.client.report [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1091.471414] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1091.471414] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1091.471414] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Deleting the datastore file [datastore1] ba8404c7-b43e-4776-941e-4d9bd8fb4c9d {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.471414] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03103a30-0113-4143-a438-eccfb51e265a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.480209] env[63293]: DEBUG oslo_vmware.api [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1091.480209] env[63293]: value = "task-1328279" [ 1091.480209] env[63293]: _type = "Task" [ 1091.480209] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.489767] env[63293]: DEBUG oslo_vmware.api [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328279, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.597842] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 76ebe2ec-332c-4d3a-957e-d0c928dafdc7] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.650914] env[63293]: DEBUG nova.compute.manager [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63293) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1091.856380] env[63293]: DEBUG nova.network.neutron [-] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.884828] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9393d584-870b-447e-abd8-09f31c69eeb8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.894814] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9fd469-6a37-4ed6-898b-d949474127bc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.920948] env[63293]: DEBUG nova.compute.manager [req-e923c9f9-641a-4599-9f4e-e8883af414b6 req-9d891425-27c6-462c-8974-c385d8b9e23f service nova] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Detach interface failed, port_id=a6a90b56-3edd-4270-a9f8-181df8dbb637, reason: Instance 97da107a-2e15-4da2-9197-7f85d960ff27 could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1091.961673] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.198s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.992537] env[63293]: DEBUG oslo_vmware.api [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328279, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146773} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.992801] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.993055] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1091.993304] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1091.993503] env[63293]: INFO nova.compute.manager [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1091.993750] env[63293]: DEBUG oslo.service.loopingcall [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.993970] env[63293]: DEBUG nova.compute.manager [-] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1091.994079] env[63293]: DEBUG nova.network.neutron [-] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1091.999782] env[63293]: INFO nova.network.neutron [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updating port 3313c3a6-17f4-4c97-b19f-9edfd6eb4099 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1092.101461] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 5f01ce51-1f5d-44ad-97f0-2306cbf55b9a] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.950685] env[63293]: INFO nova.compute.manager [-] [instance: 97da107a-2e15-4da2-9197-7f85d960ff27] Took 2.07 seconds to deallocate network for instance. [ 1092.951301] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: a15fc833-d9e8-4340-ab4f-c4a0b2d0e703] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.409533] env[63293]: DEBUG nova.compute.manager [req-861308b1-2541-459d-9e5c-73fd32cdfe01 req-d3f47ebd-269f-49b8-ba30-034f14eb9a64 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Received event network-vif-deleted-eef289b9-0d13-4ce5-be81-a5247187e457 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1093.409890] env[63293]: INFO nova.compute.manager [req-861308b1-2541-459d-9e5c-73fd32cdfe01 req-d3f47ebd-269f-49b8-ba30-034f14eb9a64 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Neutron deleted interface eef289b9-0d13-4ce5-be81-a5247187e457; detaching it from the instance and deleting it from the info cache [ 1093.409890] env[63293]: DEBUG nova.network.neutron [req-861308b1-2541-459d-9e5c-73fd32cdfe01 req-d3f47ebd-269f-49b8-ba30-034f14eb9a64 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.454650] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.454909] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.456077] env[63293]: DEBUG nova.network.neutron [-] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.457073] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 828a1165-3829-431c-9edb-3a5d3a8054ec] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.461018] env[63293]: DEBUG oslo_concurrency.lockutils [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.688892] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.689097] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.689257] env[63293]: DEBUG nova.network.neutron [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1093.912825] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb371533-8e81-43d9-9bfd-d2b8c1c40173 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.922372] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29cf8228-48ba-4593-8955-cf1169ac5216 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.947185] env[63293]: DEBUG nova.compute.manager [req-861308b1-2541-459d-9e5c-73fd32cdfe01 req-d3f47ebd-269f-49b8-ba30-034f14eb9a64 service nova] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Detach interface failed, port_id=eef289b9-0d13-4ce5-be81-a5247187e457, reason: Instance ba8404c7-b43e-4776-941e-4d9bd8fb4c9d could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1093.959205] env[63293]: INFO nova.compute.manager [-] [instance: ba8404c7-b43e-4776-941e-4d9bd8fb4c9d] Took 1.97 seconds to deallocate network for instance. [ 1093.959660] env[63293]: DEBUG nova.objects.instance [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'migration_context' on Instance uuid bea1ea16-c224-4a0e-86ed-c2aaf920d3e1 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.960815] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: d753598d-a92a-4515-9ad1-d386294f7a99] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.387361] env[63293]: DEBUG nova.network.neutron [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updating instance_info_cache with network_info: [{"id": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "address": "fa:16:3e:85:83:c3", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3313c3a6-17", "ovs_interfaceid": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.466954] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 9973dd7f-ca75-4db0-8d2b-f21b0311abeb] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.469613] env[63293]: DEBUG oslo_concurrency.lockutils [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.550318] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc6e979-5169-4a8e-bfa8-0971864a85ad {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.557568] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e169a2-2ea8-4c0b-a518-4b8809277726 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.586210] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b66ca3-0d9f-468b-a4b7-53770f8f1ca8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.592972] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370b9b08-2249-4210-beeb-1a391606c81e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.605503] env[63293]: DEBUG nova.compute.provider_tree [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.890461] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.940283] env[63293]: DEBUG nova.virt.hardware [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='3efac0c423c732ef543733f120b38e89',container_format='bare',created_at=2024-10-21T17:42:46Z,direct_url=,disk_format='vmdk',id=e4ccc7b7-9af6-4851-89da-fd23cc904ad3,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1409799387-shelved',owner='affba61d6a2846b38666544bc2c25db5',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2024-10-21T17:43:01Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1094.940545] env[63293]: DEBUG nova.virt.hardware [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1094.940707] env[63293]: DEBUG nova.virt.hardware [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1094.940924] env[63293]: DEBUG nova.virt.hardware [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1094.941106] env[63293]: DEBUG nova.virt.hardware [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1094.941262] env[63293]: DEBUG nova.virt.hardware [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1094.941488] env[63293]: DEBUG nova.virt.hardware [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1094.941664] env[63293]: DEBUG nova.virt.hardware [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1094.941842] env[63293]: DEBUG nova.virt.hardware [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1094.942068] env[63293]: DEBUG nova.virt.hardware [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1094.942259] env[63293]: DEBUG nova.virt.hardware [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.943204] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b993bbe-cad4-4eb7-89c5-a3762310ddfd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.950890] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f18104c-cea3-4a00-a638-c9f5eba770d9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.964063] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:83:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24144f5a-050a-4f1e-8d8c-774dc16dc791', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3313c3a6-17f4-4c97-b19f-9edfd6eb4099', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1094.971285] env[63293]: DEBUG oslo.service.loopingcall [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.971662] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 8ea7f3c3-8d73-4fb2-a848-4be4649e5660] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.973372] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1094.973593] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-483e6a86-c2c5-4082-b343-42150541a774 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.994720] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1094.994720] env[63293]: value = "task-1328280" [ 1094.994720] env[63293]: _type = "Task" [ 1094.994720] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.003070] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328280, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.108947] env[63293]: DEBUG nova.scheduler.client.report [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.437141] env[63293]: DEBUG nova.compute.manager [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Received event network-vif-plugged-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1095.437374] env[63293]: DEBUG oslo_concurrency.lockutils [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] Acquiring lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.437596] env[63293]: DEBUG oslo_concurrency.lockutils [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.437764] env[63293]: DEBUG oslo_concurrency.lockutils [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.437936] env[63293]: DEBUG nova.compute.manager [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] No waiting events found dispatching network-vif-plugged-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1095.438113] env[63293]: WARNING nova.compute.manager [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Received unexpected event network-vif-plugged-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 for instance with vm_state shelved_offloaded and task_state spawning. [ 1095.438273] env[63293]: DEBUG nova.compute.manager [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Received event network-changed-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1095.438427] env[63293]: DEBUG nova.compute.manager [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Refreshing instance network info cache due to event network-changed-3313c3a6-17f4-4c97-b19f-9edfd6eb4099. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1095.438611] env[63293]: DEBUG oslo_concurrency.lockutils [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] Acquiring lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.438747] env[63293]: DEBUG oslo_concurrency.lockutils [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] Acquired lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.438906] env[63293]: DEBUG nova.network.neutron [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Refreshing network info cache for port 3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1095.488049] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: f427933b-dd2c-49a9-b401-a647635a1eab] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.504629] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328280, 'name': CreateVM_Task, 'duration_secs': 0.313175} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.504629] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1095.505313] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.505313] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.505588] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1095.506136] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8d15169-96c5-414d-9cf4-58c33efd87d7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.510723] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1095.510723] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5259d6ef-5657-cff1-90f9-838e49b36547" [ 1095.510723] env[63293]: _type = "Task" [ 1095.510723] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.518010] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5259d6ef-5657-cff1-90f9-838e49b36547, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.990985] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 9ddf4f23-1279-4cbd-8212-10f344060445] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.021744] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.021958] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Processing image e4ccc7b7-9af6-4851-89da-fd23cc904ad3 {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1096.022211] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3/e4ccc7b7-9af6-4851-89da-fd23cc904ad3.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.022531] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3/e4ccc7b7-9af6-4851-89da-fd23cc904ad3.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.022695] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1096.022894] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-35c6b1ac-d1c5-4764-a65b-f377915e48ee {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.031314] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1096.031543] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1096.032240] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c60e20e6-4240-4406-b7e6-c42111725a70 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.039860] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1096.039860] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d35c58-2e20-c7ea-fe69-67879cdf3391" [ 1096.039860] env[63293]: _type = "Task" [ 1096.039860] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.047057] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52d35c58-2e20-c7ea-fe69-67879cdf3391, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.121380] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.666s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.127032] env[63293]: DEBUG oslo_concurrency.lockutils [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.666s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.127534] env[63293]: DEBUG nova.objects.instance [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lazy-loading 'resources' on Instance uuid 97da107a-2e15-4da2-9197-7f85d960ff27 {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.149835] env[63293]: DEBUG nova.network.neutron [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updated VIF entry in instance network info cache for port 3313c3a6-17f4-4c97-b19f-9edfd6eb4099. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1096.150211] env[63293]: DEBUG nova.network.neutron [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updating instance_info_cache with network_info: [{"id": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "address": "fa:16:3e:85:83:c3", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3313c3a6-17", "ovs_interfaceid": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.494761] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: c26d6a3a-3643-4b52-870a-4e2ed901d908] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.550343] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Preparing fetch location {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1096.550593] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Fetch image to [datastore1] OSTACK_IMG_47fad714-5eec-4f67-a86e-a0c56564a025/OSTACK_IMG_47fad714-5eec-4f67-a86e-a0c56564a025.vmdk {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1096.550779] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Downloading stream optimized image e4ccc7b7-9af6-4851-89da-fd23cc904ad3 to [datastore1] OSTACK_IMG_47fad714-5eec-4f67-a86e-a0c56564a025/OSTACK_IMG_47fad714-5eec-4f67-a86e-a0c56564a025.vmdk on the data store datastore1 as vApp {{(pid=63293) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1096.550951] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Downloading image file data e4ccc7b7-9af6-4851-89da-fd23cc904ad3 to the ESX as VM named 'OSTACK_IMG_47fad714-5eec-4f67-a86e-a0c56564a025' {{(pid=63293) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1096.623226] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1096.623226] env[63293]: value = "resgroup-9" [ 1096.623226] env[63293]: _type = "ResourcePool" [ 1096.623226] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1096.623551] env[63293]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0549cb44-1030-4474-b5d2-8a88371041d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.647750] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lease: (returnval){ [ 1096.647750] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f844f4-1f1a-2e15-a7ea-bc1d1c3c1d5d" [ 1096.647750] env[63293]: _type = "HttpNfcLease" [ 1096.647750] env[63293]: } obtained for vApp import into resource pool (val){ [ 1096.647750] env[63293]: value = "resgroup-9" [ 1096.647750] env[63293]: _type = "ResourcePool" [ 1096.647750] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1096.648197] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the lease: (returnval){ [ 1096.648197] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f844f4-1f1a-2e15-a7ea-bc1d1c3c1d5d" [ 1096.648197] env[63293]: _type = "HttpNfcLease" [ 1096.648197] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1096.652213] env[63293]: DEBUG oslo_concurrency.lockutils [req-9aba02d5-f3da-471d-925a-b88dc8fb1d9c req-f45fcbf4-e1a5-421c-a92b-5a40b6961b9c service nova] Releasing lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.656089] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1096.656089] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f844f4-1f1a-2e15-a7ea-bc1d1c3c1d5d" [ 1096.656089] env[63293]: _type = "HttpNfcLease" [ 1096.656089] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1096.732431] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ab05c3-8689-465f-9a4b-eaf361f6d871 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.740077] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c01492-5b86-45c9-9280-a6e72e7396f2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.769617] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7988e70-a8f3-4ac3-8b1f-ec25197e9ae2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.777203] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96e9a9b-0792-4ca1-9eef-84c20a9d0976 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.791024] env[63293]: DEBUG nova.compute.provider_tree [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1096.998488] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 93080446-fd45-4281-bf6e-35d1d964f41e] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.159291] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1097.159291] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f844f4-1f1a-2e15-a7ea-bc1d1c3c1d5d" [ 1097.159291] env[63293]: _type = "HttpNfcLease" [ 1097.159291] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1097.294641] env[63293]: DEBUG nova.scheduler.client.report [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1097.501932] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 4593cabd-1331-4ee1-99f5-46edeac56859] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.662614] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1097.662614] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f844f4-1f1a-2e15-a7ea-bc1d1c3c1d5d" [ 1097.662614] env[63293]: _type = "HttpNfcLease" [ 1097.662614] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1097.666296] env[63293]: INFO nova.compute.manager [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Swapping old allocation on dict_keys(['619725c3-6194-4724-94b1-1dd499be55d0']) held by migration eff19c51-22a9-426e-9f8b-a1e793420d8e for instance [ 1097.686107] env[63293]: DEBUG nova.scheduler.client.report [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Overwriting current allocation {'allocations': {'619725c3-6194-4724-94b1-1dd499be55d0': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 137}}, 'project_id': '1687c33183e74b6ba70e021b6879cb93', 'user_id': 'ebf82e99427d4171a4d510f7d3c966d9', 'consumer_generation': 1} on consumer bea1ea16-c224-4a0e-86ed-c2aaf920d3e1 {{(pid=63293) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1097.772658] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.772861] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.773051] env[63293]: DEBUG nova.network.neutron [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1097.799921] env[63293]: DEBUG oslo_concurrency.lockutils [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.673s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.802033] env[63293]: DEBUG oslo_concurrency.lockutils [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.332s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.802317] env[63293]: DEBUG nova.objects.instance [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lazy-loading 'resources' on Instance uuid ba8404c7-b43e-4776-941e-4d9bd8fb4c9d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.817821] env[63293]: INFO nova.scheduler.client.report [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleted allocations for instance 97da107a-2e15-4da2-9197-7f85d960ff27 [ 1098.023678] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 1e01a75d-122d-4122-9fed-4164d64d4ee8] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.161915] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1098.161915] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f844f4-1f1a-2e15-a7ea-bc1d1c3c1d5d" [ 1098.161915] env[63293]: _type = "HttpNfcLease" [ 1098.161915] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1098.325647] env[63293]: DEBUG oslo_concurrency.lockutils [None req-41db9953-1047-4e70-bb8f-4ce32bb30a5b tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "97da107a-2e15-4da2-9197-7f85d960ff27" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.591s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.359458] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0bd9cf-d1e7-4bcc-97ae-6e044bc76c18 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.366768] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0624a3-ac7d-453a-95d3-5a83c63b33aa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.398638] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775d3fa3-1556-44df-adc7-a7c0efdea723 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.405881] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6b384b-e0eb-463b-98ab-36ea686c8fcc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.418473] env[63293]: DEBUG nova.compute.provider_tree [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.508224] env[63293]: DEBUG nova.network.neutron [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance_info_cache with network_info: [{"id": "c589de77-25fc-43f7-9480-e6c97bbda93d", "address": "fa:16:3e:4a:ef:d2", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc589de77-25", "ovs_interfaceid": "c589de77-25fc-43f7-9480-e6c97bbda93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.526019] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 831d7d9b-364f-4c29-bda5-e6c38291a973] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.662596] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1098.662596] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f844f4-1f1a-2e15-a7ea-bc1d1c3c1d5d" [ 1098.662596] env[63293]: _type = "HttpNfcLease" [ 1098.662596] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1098.662857] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1098.662857] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52f844f4-1f1a-2e15-a7ea-bc1d1c3c1d5d" [ 1098.662857] env[63293]: _type = "HttpNfcLease" [ 1098.662857] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1098.663552] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa201ad0-224e-4a64-881b-9e9bbe150b37 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.670273] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52450def-2a38-e194-7330-f304d502a3f6/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1098.670450] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52450def-2a38-e194-7330-f304d502a3f6/disk-0.vmdk. {{(pid=63293) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1098.739145] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2d1520df-4dfb-48ef-bfc2-bb65e19c6772 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.923037] env[63293]: DEBUG nova.scheduler.client.report [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.010613] env[63293]: DEBUG oslo_concurrency.lockutils [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.012185] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a88c7ea-7581-4816-8507-35e505ef173b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.023341] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5915f8-653a-4196-bd48-b112b3444d8c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.030820] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: ad585ebb-2072-45df-b645-94c9fa93576b] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.429071] env[63293]: DEBUG oslo_concurrency.lockutils [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.627s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.449979] env[63293]: INFO nova.scheduler.client.report [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Deleted allocations for instance ba8404c7-b43e-4776-941e-4d9bd8fb4c9d [ 1099.533578] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 48603812-c3cc-4dae-b8ba-99b9ac5f7969] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.801205] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Completed reading data from the image iterator. {{(pid=63293) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1099.801457] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52450def-2a38-e194-7330-f304d502a3f6/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1099.802516] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970fdb1f-f563-46e5-89ee-c3ec5dce54b1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.809014] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52450def-2a38-e194-7330-f304d502a3f6/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1099.809190] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52450def-2a38-e194-7330-f304d502a3f6/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1099.809420] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-1629b482-2eb3-4438-a2ca-9d0de15c3c64 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.957371] env[63293]: DEBUG oslo_concurrency.lockutils [None req-57910d91-1966-44ea-bd44-8c8796c9cfbd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "ba8404c7-b43e-4776-941e-4d9bd8fb4c9d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.114s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.037298] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: b8637881-9fb6-4c5d-848d-7b2d38f8e970] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.048887] env[63293]: DEBUG oslo_vmware.rw_handles [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52450def-2a38-e194-7330-f304d502a3f6/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1100.048887] env[63293]: INFO nova.virt.vmwareapi.images [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Downloaded image file data e4ccc7b7-9af6-4851-89da-fd23cc904ad3 [ 1100.049405] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d8cd68-d3af-4358-81f0-e527fdbd57a6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.068835] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e32b6a8-3cd1-45eb-b5e7-fec6e760b440 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.104804] env[63293]: INFO nova.virt.vmwareapi.images [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] The imported VM was unregistered [ 1100.107178] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Caching image {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1100.108809] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Creating directory with path [datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3 {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1100.108809] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f8f6169-85d5-4c01-9ee1-a1926b8b9859 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.118835] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Created directory with path [datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3 {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1100.118957] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_47fad714-5eec-4f67-a86e-a0c56564a025/OSTACK_IMG_47fad714-5eec-4f67-a86e-a0c56564a025.vmdk to [datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3/e4ccc7b7-9af6-4851-89da-fd23cc904ad3.vmdk. {{(pid=63293) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1100.119209] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-7cd701ad-0449-4ff0-a78a-d4131a32d5e5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.126912] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1100.126912] env[63293]: value = "task-1328283" [ 1100.126912] env[63293]: _type = "Task" [ 1100.126912] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.130707] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1100.131450] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-533c36b0-09d6-41be-99e1-64418d58e237 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.137568] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328283, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.138824] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1100.138824] env[63293]: value = "task-1328284" [ 1100.138824] env[63293]: _type = "Task" [ 1100.138824] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.147878] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328284, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.541648] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 33d71260-26f1-482c-b93b-5f8e27c545f3] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.616434] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "07418025-09fc-4793-a8ea-f059f42251ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.616763] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "07418025-09fc-4793-a8ea-f059f42251ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.617017] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "07418025-09fc-4793-a8ea-f059f42251ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.617224] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "07418025-09fc-4793-a8ea-f059f42251ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.617404] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "07418025-09fc-4793-a8ea-f059f42251ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.619698] env[63293]: INFO nova.compute.manager [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Terminating instance [ 1100.621703] env[63293]: DEBUG nova.compute.manager [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1100.621920] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1100.622760] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f8ea54-cd2e-4041-9c32-ea65276dc2f1 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.633445] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1100.634098] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-92ef1ae7-1665-4865-8746-b555abaa5fe5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.638546] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328283, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.643933] env[63293]: DEBUG oslo_vmware.api [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1100.643933] env[63293]: value = "task-1328285" [ 1100.643933] env[63293]: _type = "Task" [ 1100.643933] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.647388] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328284, 'name': PowerOffVM_Task, 'duration_secs': 0.234287} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.650493] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1100.651218] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1100.651506] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1100.651704] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1100.651929] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1100.652103] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1100.652323] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1100.652580] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1100.652795] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1100.652978] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1100.653166] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1100.653348] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1100.658538] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d88eecf2-0d4f-46f8-a8b1-0fded18dd277 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.674813] env[63293]: DEBUG oslo_vmware.api [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328285, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.676362] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1100.676362] env[63293]: value = "task-1328286" [ 1100.676362] env[63293]: _type = "Task" [ 1100.676362] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.684221] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328286, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.045548] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 65c2f72d-6bbe-45ac-8efc-401e0664390f] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.142018] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.142282] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.143465] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328283, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.158288] env[63293]: DEBUG oslo_vmware.api [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328285, 'name': PowerOffVM_Task, 'duration_secs': 0.246584} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.158569] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1101.158742] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1101.159009] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1493636e-046e-4d5e-b625-3c5a27868e9b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.186464] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328286, 'name': ReconfigVM_Task, 'duration_secs': 0.324191} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.187345] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eccaffad-80ad-4249-bf91-63d0e3979505 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.210203] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1101.210638] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1101.210828] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1101.211039] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1101.211279] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1101.211533] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1101.211877] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1101.212100] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1101.212304] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1101.212491] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1101.212710] env[63293]: DEBUG nova.virt.hardware [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1101.214057] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e720797-a400-49ba-a480-284d1a3bf81f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.220068] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1101.220068] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52755533-c7d9-aaad-8f84-b2b4b91ce565" [ 1101.220068] env[63293]: _type = "Task" [ 1101.220068] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.229134] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52755533-c7d9-aaad-8f84-b2b4b91ce565, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.232667] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1101.232929] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1101.233143] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Deleting the datastore file [datastore2] 07418025-09fc-4793-a8ea-f059f42251ea {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1101.233415] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4baa0df3-757a-487a-9bf0-7463133ceb7b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.239394] env[63293]: DEBUG oslo_vmware.api [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for the task: (returnval){ [ 1101.239394] env[63293]: value = "task-1328288" [ 1101.239394] env[63293]: _type = "Task" [ 1101.239394] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.247512] env[63293]: DEBUG oslo_vmware.api [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328288, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.549574] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: c8009fc0-f7dc-4568-9909-7fa9f7f6a56c] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.638762] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328283, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.644828] env[63293]: DEBUG nova.compute.manager [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1101.730282] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52755533-c7d9-aaad-8f84-b2b4b91ce565, 'name': SearchDatastore_Task, 'duration_secs': 0.015449} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.737468] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfiguring VM instance instance-00000064 to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1101.738026] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aed3c6de-a595-4711-88b6-3cdeb5fbacc9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.763162] env[63293]: DEBUG oslo_vmware.api [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328288, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.765630] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1101.765630] env[63293]: value = "task-1328289" [ 1101.765630] env[63293]: _type = "Task" [ 1101.765630] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.773626] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328289, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.056158] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 0977c79a-2ef3-4ec9-a0a0-de1ed5799723] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.140459] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328283, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.169840] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.170126] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.171790] env[63293]: INFO nova.compute.claims [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1102.261113] env[63293]: DEBUG oslo_vmware.api [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Task: {'id': task-1328288, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.633238} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.261397] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1102.261605] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1102.261781] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1102.261961] env[63293]: INFO nova.compute.manager [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1102.262227] env[63293]: DEBUG oslo.service.loopingcall [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1102.262428] env[63293]: DEBUG nova.compute.manager [-] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1102.262523] env[63293]: DEBUG nova.network.neutron [-] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1102.275565] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.529575] env[63293]: DEBUG nova.compute.manager [req-ca5bc52a-28ff-4f30-a2f2-4746ec66fdcd req-50775bd7-ef0a-43d4-aa2e-043a04ccff36 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Received event network-vif-deleted-de6fe4cc-5e56-4066-bee8-f16b1fd6bed4 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1102.529731] env[63293]: INFO nova.compute.manager [req-ca5bc52a-28ff-4f30-a2f2-4746ec66fdcd req-50775bd7-ef0a-43d4-aa2e-043a04ccff36 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Neutron deleted interface de6fe4cc-5e56-4066-bee8-f16b1fd6bed4; detaching it from the instance and deleting it from the info cache [ 1102.529909] env[63293]: DEBUG nova.network.neutron [req-ca5bc52a-28ff-4f30-a2f2-4746ec66fdcd req-50775bd7-ef0a-43d4-aa2e-043a04ccff36 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.559586] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: cc77e4f6-8e2f-4f2c-a1b9-dc5ee2096395] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.639601] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328283, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.776296] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328289, 'name': ReconfigVM_Task, 'duration_secs': 0.609709} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.776604] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfigured VM instance instance-00000064 to detach disk 2000 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1102.777376] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd2ca80-07e5-441c-aef3-9a1c42926d43 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.802115] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] bea1ea16-c224-4a0e-86ed-c2aaf920d3e1/bea1ea16-c224-4a0e-86ed-c2aaf920d3e1.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1102.802372] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ae7d3e8-9b1a-47a1-9444-c184140d18e4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.820830] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1102.820830] env[63293]: value = "task-1328290" [ 1102.820830] env[63293]: _type = "Task" [ 1102.820830] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.833423] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328290, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.011648] env[63293]: DEBUG nova.network.neutron [-] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.031953] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d6b2c059-0d6c-4b5d-8563-826f60fff4b8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.043057] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0bf303-482a-45c3-92e7-5c70f30c8f05 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.062116] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 495cb1b0-9194-44e6-9f68-de62033ea17d] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.068696] env[63293]: DEBUG nova.compute.manager [req-ca5bc52a-28ff-4f30-a2f2-4746ec66fdcd req-50775bd7-ef0a-43d4-aa2e-043a04ccff36 service nova] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Detach interface failed, port_id=de6fe4cc-5e56-4066-bee8-f16b1fd6bed4, reason: Instance 07418025-09fc-4793-a8ea-f059f42251ea could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1103.141927] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328283, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.645084} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.142239] env[63293]: INFO nova.virt.vmwareapi.ds_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_47fad714-5eec-4f67-a86e-a0c56564a025/OSTACK_IMG_47fad714-5eec-4f67-a86e-a0c56564a025.vmdk to [datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3/e4ccc7b7-9af6-4851-89da-fd23cc904ad3.vmdk. [ 1103.142444] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Cleaning up location [datastore1] OSTACK_IMG_47fad714-5eec-4f67-a86e-a0c56564a025 {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1103.142626] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_47fad714-5eec-4f67-a86e-a0c56564a025 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1103.142907] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac129cfc-0531-4b82-8b38-174b35aef91a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.149241] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1103.149241] env[63293]: value = "task-1328291" [ 1103.149241] env[63293]: _type = "Task" [ 1103.149241] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.157174] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.237413] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85953e00-0b64-4665-bd67-8248431f0049 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.245103] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5004f291-2b4a-408c-9bbf-f359bae508f8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.273718] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584d763a-c53d-4aea-948d-75b906f9620f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.280423] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2355740-7637-4748-837f-734467f4739d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.293779] env[63293]: DEBUG nova.compute.provider_tree [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.329016] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328290, 'name': ReconfigVM_Task, 'duration_secs': 0.318384} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.329302] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfigured VM instance instance-00000064 to attach disk [datastore2] bea1ea16-c224-4a0e-86ed-c2aaf920d3e1/bea1ea16-c224-4a0e-86ed-c2aaf920d3e1.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1103.330089] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa128b4-4c37-4d6c-a792-ceac8191df88 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.350263] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a66ab4-0fca-4d17-ba20-166865cd9cd8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.370100] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cc743a-2e0f-4e19-8098-153ff1d21557 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.389786] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310f5f79-820b-4a08-be7e-6ac3d45fae49 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.396174] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1103.396387] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26736b5f-c5e1-4a9a-8aa7-9335cff55cdf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.401472] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1103.401472] env[63293]: value = "task-1328292" [ 1103.401472] env[63293]: _type = "Task" [ 1103.401472] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.408433] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328292, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.514619] env[63293]: INFO nova.compute.manager [-] [instance: 07418025-09fc-4793-a8ea-f059f42251ea] Took 1.25 seconds to deallocate network for instance. [ 1103.570244] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: b48324a4-902d-4642-8827-daee9683b3a0] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.658808] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.039571} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.659092] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1103.659271] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3/e4ccc7b7-9af6-4851-89da-fd23cc904ad3.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.659515] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3/e4ccc7b7-9af6-4851-89da-fd23cc904ad3.vmdk to [datastore1] 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d/97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1103.659793] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9aec9be4-0061-4fed-af7f-3fec99b3aa6a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.668260] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1103.668260] env[63293]: value = "task-1328293" [ 1103.668260] env[63293]: _type = "Task" [ 1103.668260] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.675600] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328293, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.796543] env[63293]: DEBUG nova.scheduler.client.report [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1103.913070] env[63293]: DEBUG oslo_vmware.api [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328292, 'name': PowerOnVM_Task, 'duration_secs': 0.40976} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.913349] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1104.020596] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.073820] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: b5334832-25dc-4b03-9689-388a1382200a] Instance has had 0 of 5 cleanup attempts {{(pid=63293) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.179913] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328293, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.303773] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.133s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.304354] env[63293]: DEBUG nova.compute.manager [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1104.307291] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.287s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.307525] env[63293]: DEBUG nova.objects.instance [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lazy-loading 'resources' on Instance uuid 07418025-09fc-4793-a8ea-f059f42251ea {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.680212] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328293, 'name': CopyVirtualDisk_Task} progress is 29%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.810656] env[63293]: DEBUG nova.compute.utils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1104.812217] env[63293]: DEBUG nova.compute.manager [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1104.812395] env[63293]: DEBUG nova.network.neutron [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1104.866476] env[63293]: DEBUG nova.policy [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7a950baa92f4d9ea77e5d6461582e0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec53642df3804e7190615487a426d4a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 1104.882604] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c369e23a-02d7-4f3e-8f30-af3842a76538 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.891950] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ce86dd-d753-4cf8-b595-e7ff93c7ca27 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.928580] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7d2463-a515-4221-a5f6-208ae12e5a33 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.937232] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f38bb53-97b9-409a-a415-91036c093cb6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.954092] env[63293]: DEBUG nova.compute.provider_tree [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.963681] env[63293]: INFO nova.compute.manager [None req-b0eb44af-559b-4411-9f4d-6766f34655bd tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance to original state: 'active' [ 1105.182934] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328293, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.234689] env[63293]: DEBUG nova.network.neutron [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Successfully created port: 5e70c362-e59e-422f-b812-98271464078c {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1105.318415] env[63293]: DEBUG nova.compute.manager [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1105.461023] env[63293]: DEBUG nova.scheduler.client.report [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1105.681169] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328293, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.964608] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.657s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.991269] env[63293]: INFO nova.scheduler.client.report [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Deleted allocations for instance 07418025-09fc-4793-a8ea-f059f42251ea [ 1106.182790] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328293, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.328339] env[63293]: DEBUG nova.compute.manager [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1106.353582] env[63293]: DEBUG nova.virt.hardware [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1106.353825] env[63293]: DEBUG nova.virt.hardware [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1106.353996] env[63293]: DEBUG nova.virt.hardware [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1106.354288] env[63293]: DEBUG nova.virt.hardware [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1106.354465] env[63293]: DEBUG nova.virt.hardware [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1106.354625] env[63293]: DEBUG nova.virt.hardware [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1106.354837] env[63293]: DEBUG nova.virt.hardware [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1106.355008] env[63293]: DEBUG nova.virt.hardware [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1106.355189] env[63293]: DEBUG nova.virt.hardware [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1106.355356] env[63293]: DEBUG nova.virt.hardware [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1106.355536] env[63293]: DEBUG nova.virt.hardware [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1106.356501] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81bdbe3-40cf-4fe1-9b2e-82ce34e76444 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.364544] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724d7a00-3e1d-4efd-8469-79a6e8dcd780 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.469111] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.500639] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4eeb0a26-3277-4f06-9fdf-e0dd9b148cdd tempest-ServerRescueTestJSON-829158820 tempest-ServerRescueTestJSON-829158820-project-member] Lock "07418025-09fc-4793-a8ea-f059f42251ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.884s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.662969] env[63293]: DEBUG nova.compute.manager [req-be608286-e453-4bfc-acd4-e2792f36abd7 req-991037ee-3ca7-4a5f-b668-8336b395e7a5 service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Received event network-vif-plugged-5e70c362-e59e-422f-b812-98271464078c {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1106.663190] env[63293]: DEBUG oslo_concurrency.lockutils [req-be608286-e453-4bfc-acd4-e2792f36abd7 req-991037ee-3ca7-4a5f-b668-8336b395e7a5 service nova] Acquiring lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.663378] env[63293]: DEBUG oslo_concurrency.lockutils [req-be608286-e453-4bfc-acd4-e2792f36abd7 req-991037ee-3ca7-4a5f-b668-8336b395e7a5 service nova] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.663496] env[63293]: DEBUG oslo_concurrency.lockutils [req-be608286-e453-4bfc-acd4-e2792f36abd7 req-991037ee-3ca7-4a5f-b668-8336b395e7a5 service nova] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.663672] env[63293]: DEBUG nova.compute.manager [req-be608286-e453-4bfc-acd4-e2792f36abd7 req-991037ee-3ca7-4a5f-b668-8336b395e7a5 service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] No waiting events found dispatching network-vif-plugged-5e70c362-e59e-422f-b812-98271464078c {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1106.663842] env[63293]: WARNING nova.compute.manager [req-be608286-e453-4bfc-acd4-e2792f36abd7 req-991037ee-3ca7-4a5f-b668-8336b395e7a5 service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Received unexpected event network-vif-plugged-5e70c362-e59e-422f-b812-98271464078c for instance with vm_state building and task_state spawning. [ 1106.682198] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328293, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.616333} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.682471] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e4ccc7b7-9af6-4851-89da-fd23cc904ad3/e4ccc7b7-9af6-4851-89da-fd23cc904ad3.vmdk to [datastore1] 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d/97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1106.683206] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e1170d-a5a0-423a-a70e-bdc0c351fcf4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.704747] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d/97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d.vmdk or device None with type streamOptimized {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.705306] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57b14688-8fb3-466e-912e-04a76676353f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.727434] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1106.727434] env[63293]: value = "task-1328294" [ 1106.727434] env[63293]: _type = "Task" [ 1106.727434] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.735907] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328294, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.752895] env[63293]: DEBUG nova.network.neutron [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Successfully updated port: 5e70c362-e59e-422f-b812-98271464078c {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1106.931988] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.932451] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.932658] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.932882] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.933077] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.935241] env[63293]: INFO nova.compute.manager [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Terminating instance [ 1106.937140] env[63293]: DEBUG nova.compute.manager [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1106.937360] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1106.937596] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0aeb89f-4b25-477b-954e-0dfdb4e17ac4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.946292] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1106.946292] env[63293]: value = "task-1328295" [ 1106.946292] env[63293]: _type = "Task" [ 1106.946292] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.955316] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328295, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.975898] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._sync_power_states {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.238169] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328294, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.258269] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "refresh_cache-fa68f69e-febd-4a46-81b8-afb7e5ce048d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.258431] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "refresh_cache-fa68f69e-febd-4a46-81b8-afb7e5ce048d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.258546] env[63293]: DEBUG nova.network.neutron [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1107.459617] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328295, 'name': PowerOffVM_Task, 'duration_secs': 0.200886} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.459901] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1107.460777] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Volume detach. Driver type: vmdk {{(pid=63293) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1107.461058] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283845', 'volume_id': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'name': 'volume-c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'bea1ea16-c224-4a0e-86ed-c2aaf920d3e1', 'attached_at': '2024-10-21T17:43:22.000000', 'detached_at': '', 'volume_id': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'serial': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1107.462525] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f32bf2-e7f4-4aa7-903a-20a5a79ce848 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.484313] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Getting list of instances from cluster (obj){ [ 1107.484313] env[63293]: value = "domain-c8" [ 1107.484313] env[63293]: _type = "ClusterComputeResource" [ 1107.484313] env[63293]: } {{(pid=63293) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1107.485438] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5918b6c-f5f7-4212-8458-a01ca6f449b2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.489414] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3f284d-28db-4fbe-9b90-6253b42b743a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.501548] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e6ea7f-f114-4875-8a0c-6b3424b0cad9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.504704] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Got total of 2 instances {{(pid=63293) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1107.504864] env[63293]: WARNING nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] While synchronizing instance power states, found 3 instances in the database and 2 instances on the hypervisor. [ 1107.505008] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Triggering sync for uuid bea1ea16-c224-4a0e-86ed-c2aaf920d3e1 {{(pid=63293) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1107.505203] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Triggering sync for uuid 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d {{(pid=63293) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1107.505543] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Triggering sync for uuid fa68f69e-febd-4a46-81b8-afb7e5ce048d {{(pid=63293) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1107.505866] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.506122] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.506335] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.525015] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ac43e1-a179-4ad8-b786-b756f03dee49 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.539475] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] The volume has not been displaced from its original location: [datastore1] volume-c4a57846-fa87-44ef-a7e0-0cd64f5a6037/volume-c4a57846-fa87-44ef-a7e0-0cd64f5a6037.vmdk. No consolidation needed. {{(pid=63293) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1107.544679] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfiguring VM instance instance-00000064 to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1107.545027] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89c0763c-9513-4312-b947-abcb85132b42 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.563570] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1107.563570] env[63293]: value = "task-1328296" [ 1107.563570] env[63293]: _type = "Task" [ 1107.563570] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.570965] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328296, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.737805] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328294, 'name': ReconfigVM_Task, 'duration_secs': 0.588953} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.738239] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d/97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d.vmdk or device None with type streamOptimized {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1107.738805] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a89e8be6-930b-4974-9ae1-ffb5e67db8d5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.745706] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1107.745706] env[63293]: value = "task-1328297" [ 1107.745706] env[63293]: _type = "Task" [ 1107.745706] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.753492] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328297, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.795438] env[63293]: DEBUG nova.network.neutron [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1108.075780] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328296, 'name': ReconfigVM_Task, 'duration_secs': 0.247572} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.076105] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Reconfigured VM instance instance-00000064 to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1108.080746] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1078a2b8-8f20-4272-90a3-353fc8d6c976 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.096194] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1108.096194] env[63293]: value = "task-1328298" [ 1108.096194] env[63293]: _type = "Task" [ 1108.096194] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.105446] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328298, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.180245] env[63293]: DEBUG nova.network.neutron [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Updating instance_info_cache with network_info: [{"id": "5e70c362-e59e-422f-b812-98271464078c", "address": "fa:16:3e:74:52:dc", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e70c362-e5", "ovs_interfaceid": "5e70c362-e59e-422f-b812-98271464078c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.255287] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328297, 'name': Rename_Task, 'duration_secs': 0.144206} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.255579] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1108.255830] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94976b2d-4803-40d7-9e13-ecf03c2e0fb4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.262193] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1108.262193] env[63293]: value = "task-1328299" [ 1108.262193] env[63293]: _type = "Task" [ 1108.262193] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.271175] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328299, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.606673] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328298, 'name': ReconfigVM_Task, 'duration_secs': 0.133321} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.607100] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283845', 'volume_id': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'name': 'volume-c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'bea1ea16-c224-4a0e-86ed-c2aaf920d3e1', 'attached_at': '2024-10-21T17:43:22.000000', 'detached_at': '', 'volume_id': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037', 'serial': 'c4a57846-fa87-44ef-a7e0-0cd64f5a6037'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1108.607507] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1108.608340] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05dd24d-cdc3-447c-b112-8d3af4380be4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.614788] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1108.615033] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-76b7a94b-c1c1-4e41-9686-a7cfad1bf45d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.682435] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1108.682670] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1108.682857] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleting the datastore file [datastore2] bea1ea16-c224-4a0e-86ed-c2aaf920d3e1 {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1108.683354] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "refresh_cache-fa68f69e-febd-4a46-81b8-afb7e5ce048d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.685232] env[63293]: DEBUG nova.compute.manager [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Instance network_info: |[{"id": "5e70c362-e59e-422f-b812-98271464078c", "address": "fa:16:3e:74:52:dc", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e70c362-e5", "ovs_interfaceid": "5e70c362-e59e-422f-b812-98271464078c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1108.685232] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-370faa45-11c6-40a8-9b34-78dd69d07388 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.686032] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:52:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91c1da19-ab68-4127-bacd-accbaff19651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5e70c362-e59e-422f-b812-98271464078c', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1108.694096] env[63293]: DEBUG oslo.service.loopingcall [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1108.695603] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1108.697134] env[63293]: DEBUG nova.compute.manager [req-5c346cd2-c58b-47f4-a252-9d2500e5f8d8 req-c7989fc0-720c-4607-944a-b63ff02ecffa service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Received event network-changed-5e70c362-e59e-422f-b812-98271464078c {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1108.697336] env[63293]: DEBUG nova.compute.manager [req-5c346cd2-c58b-47f4-a252-9d2500e5f8d8 req-c7989fc0-720c-4607-944a-b63ff02ecffa service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Refreshing instance network info cache due to event network-changed-5e70c362-e59e-422f-b812-98271464078c. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1108.697552] env[63293]: DEBUG oslo_concurrency.lockutils [req-5c346cd2-c58b-47f4-a252-9d2500e5f8d8 req-c7989fc0-720c-4607-944a-b63ff02ecffa service nova] Acquiring lock "refresh_cache-fa68f69e-febd-4a46-81b8-afb7e5ce048d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.697700] env[63293]: DEBUG oslo_concurrency.lockutils [req-5c346cd2-c58b-47f4-a252-9d2500e5f8d8 req-c7989fc0-720c-4607-944a-b63ff02ecffa service nova] Acquired lock "refresh_cache-fa68f69e-febd-4a46-81b8-afb7e5ce048d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.697865] env[63293]: DEBUG nova.network.neutron [req-5c346cd2-c58b-47f4-a252-9d2500e5f8d8 req-c7989fc0-720c-4607-944a-b63ff02ecffa service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Refreshing network info cache for port 5e70c362-e59e-422f-b812-98271464078c {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1108.698948] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d2e43644-590a-4e81-8a7e-e26ff176a41c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.716902] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1108.716902] env[63293]: value = "task-1328301" [ 1108.716902] env[63293]: _type = "Task" [ 1108.716902] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.722442] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1108.722442] env[63293]: value = "task-1328302" [ 1108.722442] env[63293]: _type = "Task" [ 1108.722442] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.731970] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328301, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.735185] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328302, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.773209] env[63293]: DEBUG oslo_vmware.api [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328299, 'name': PowerOnVM_Task, 'duration_secs': 0.503549} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.773535] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1108.897462] env[63293]: DEBUG nova.compute.manager [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1108.898443] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fcbf45-6773-4666-8156-8b90406b2501 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.229875] env[63293]: DEBUG oslo_vmware.api [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328301, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.344494} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.230781] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1109.231041] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1109.231186] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1109.231410] env[63293]: INFO nova.compute.manager [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Took 2.29 seconds to destroy the instance on the hypervisor. [ 1109.231620] env[63293]: DEBUG oslo.service.loopingcall [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.231832] env[63293]: DEBUG nova.compute.manager [-] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1109.231933] env[63293]: DEBUG nova.network.neutron [-] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1109.236897] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328302, 'name': CreateVM_Task} progress is 25%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.420327] env[63293]: DEBUG oslo_concurrency.lockutils [None req-c670e6a6-efaa-48b3-85b2-de716d9d1c32 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.688s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.421247] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.915s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.421450] env[63293]: INFO nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] During sync_power_state the instance has a pending task (spawning). Skip. [ 1109.421627] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.600545] env[63293]: DEBUG nova.network.neutron [req-5c346cd2-c58b-47f4-a252-9d2500e5f8d8 req-c7989fc0-720c-4607-944a-b63ff02ecffa service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Updated VIF entry in instance network info cache for port 5e70c362-e59e-422f-b812-98271464078c. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1109.600997] env[63293]: DEBUG nova.network.neutron [req-5c346cd2-c58b-47f4-a252-9d2500e5f8d8 req-c7989fc0-720c-4607-944a-b63ff02ecffa service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Updating instance_info_cache with network_info: [{"id": "5e70c362-e59e-422f-b812-98271464078c", "address": "fa:16:3e:74:52:dc", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e70c362-e5", "ovs_interfaceid": "5e70c362-e59e-422f-b812-98271464078c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.733354] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328302, 'name': CreateVM_Task, 'duration_secs': 0.822886} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.733512] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1109.734250] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.734494] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.734857] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1109.735140] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19a24964-7132-4163-ad41-417711f2e426 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.739890] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1109.739890] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5234edd9-8f1a-0249-7caf-70e50af7a522" [ 1109.739890] env[63293]: _type = "Task" [ 1109.739890] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.749842] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5234edd9-8f1a-0249-7caf-70e50af7a522, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.104200] env[63293]: DEBUG oslo_concurrency.lockutils [req-5c346cd2-c58b-47f4-a252-9d2500e5f8d8 req-c7989fc0-720c-4607-944a-b63ff02ecffa service nova] Releasing lock "refresh_cache-fa68f69e-febd-4a46-81b8-afb7e5ce048d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.250527] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5234edd9-8f1a-0249-7caf-70e50af7a522, 'name': SearchDatastore_Task, 'duration_secs': 0.010141} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.250842] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.251095] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1110.251336] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.251488] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.251694] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1110.251980] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1585b203-2ff4-4c91-b2f3-2497b7ebbd8c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.264129] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1110.264319] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1110.265016] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fae51c3-1ca2-4f66-bd25-05a9c2d12fb4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.269575] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1110.269575] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52801ae8-4464-e99e-c109-799a1a76a458" [ 1110.269575] env[63293]: _type = "Task" [ 1110.269575] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.276722] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52801ae8-4464-e99e-c109-799a1a76a458, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.328195] env[63293]: DEBUG nova.network.neutron [-] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.724396] env[63293]: DEBUG nova.compute.manager [req-68bcba16-f445-448f-b585-cb05210ab198 req-7715336f-8858-49f6-9d51-12805d9df0ed service nova] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Received event network-vif-deleted-c589de77-25fc-43f7-9480-e6c97bbda93d {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1110.779517] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52801ae8-4464-e99e-c109-799a1a76a458, 'name': SearchDatastore_Task, 'duration_secs': 0.040874} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.780506] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-969734f0-f846-4ad7-adc6-8191dbddc454 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.785813] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1110.785813] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529a59c1-a1e2-070f-8530-d47d95d65621" [ 1110.785813] env[63293]: _type = "Task" [ 1110.785813] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.793623] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529a59c1-a1e2-070f-8530-d47d95d65621, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.830234] env[63293]: INFO nova.compute.manager [-] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Took 1.60 seconds to deallocate network for instance. [ 1111.296320] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]529a59c1-a1e2-070f-8530-d47d95d65621, 'name': SearchDatastore_Task, 'duration_secs': 0.037383} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.296756] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.296892] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] fa68f69e-febd-4a46-81b8-afb7e5ce048d/fa68f69e-febd-4a46-81b8-afb7e5ce048d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1111.297176] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7337e0f7-07dc-41fd-8ab8-6e5fdc0201fc {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.303934] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1111.303934] env[63293]: value = "task-1328303" [ 1111.303934] env[63293]: _type = "Task" [ 1111.303934] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.311333] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328303, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.377326] env[63293]: INFO nova.compute.manager [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] Took 0.55 seconds to detach 1 volumes for instance. [ 1111.815144] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328303, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49379} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.815490] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] fa68f69e-febd-4a46-81b8-afb7e5ce048d/fa68f69e-febd-4a46-81b8-afb7e5ce048d.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1111.815673] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1111.815927] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-943ab3b4-b6b9-4736-b50d-ca75352f90af {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.823260] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1111.823260] env[63293]: value = "task-1328304" [ 1111.823260] env[63293]: _type = "Task" [ 1111.823260] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.830403] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328304, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.883895] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.884247] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.884506] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.910872] env[63293]: INFO nova.scheduler.client.report [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleted allocations for instance bea1ea16-c224-4a0e-86ed-c2aaf920d3e1 [ 1112.332944] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328304, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06512} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.333298] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1112.333987] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e9d584-852c-4dac-9d7a-4c145b19c139 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.355867] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] fa68f69e-febd-4a46-81b8-afb7e5ce048d/fa68f69e-febd-4a46-81b8-afb7e5ce048d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.356116] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78657a62-864a-4f14-b612-a7eff21a637b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.375424] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1112.375424] env[63293]: value = "task-1328305" [ 1112.375424] env[63293]: _type = "Task" [ 1112.375424] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.382463] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328305, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.417810] env[63293]: DEBUG oslo_concurrency.lockutils [None req-e5bc7a18-362c-4d11-9a6f-b8d5f03f6554 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.485s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.418728] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.913s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.418922] env[63293]: INFO nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: bea1ea16-c224-4a0e-86ed-c2aaf920d3e1] During sync_power_state the instance has a pending task (deleting). Skip. [ 1112.419109] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "bea1ea16-c224-4a0e-86ed-c2aaf920d3e1" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.886205] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328305, 'name': ReconfigVM_Task, 'duration_secs': 0.290707} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.886505] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Reconfigured VM instance instance-0000006d to attach disk [datastore2] fa68f69e-febd-4a46-81b8-afb7e5ce048d/fa68f69e-febd-4a46-81b8-afb7e5ce048d.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1112.887170] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95665a84-b3ef-4be7-b936-04465d79b5cb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.894063] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1112.894063] env[63293]: value = "task-1328306" [ 1112.894063] env[63293]: _type = "Task" [ 1112.894063] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.901417] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328306, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.403863] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328306, 'name': Rename_Task, 'duration_secs': 0.147861} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.404225] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1113.404418] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ddaf683-5ce1-4024-966f-a4e87488dcf7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.410677] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1113.410677] env[63293]: value = "task-1328307" [ 1113.410677] env[63293]: _type = "Task" [ 1113.410677] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.418049] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328307, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.921243] env[63293]: DEBUG oslo_vmware.api [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328307, 'name': PowerOnVM_Task, 'duration_secs': 0.440837} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.921478] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1113.921682] env[63293]: INFO nova.compute.manager [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Took 7.59 seconds to spawn the instance on the hypervisor. [ 1113.921893] env[63293]: DEBUG nova.compute.manager [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1113.922671] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa3e033-743e-4634-9d1c-cc8a59aa99a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.079908] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "7ae46382-8372-4a33-a9b6-e0720f09811e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.080160] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.438916] env[63293]: INFO nova.compute.manager [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Took 12.29 seconds to build instance. [ 1114.581969] env[63293]: DEBUG nova.compute.manager [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Starting instance... {{(pid=63293) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1114.663364] env[63293]: DEBUG nova.compute.manager [req-a8cf6bda-7c4c-43a3-bfd8-dd44522c85ed req-dcd11307-a83a-4212-9373-7f8e9fef5dfe service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Received event network-changed-5e70c362-e59e-422f-b812-98271464078c {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1114.663586] env[63293]: DEBUG nova.compute.manager [req-a8cf6bda-7c4c-43a3-bfd8-dd44522c85ed req-dcd11307-a83a-4212-9373-7f8e9fef5dfe service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Refreshing instance network info cache due to event network-changed-5e70c362-e59e-422f-b812-98271464078c. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1114.664132] env[63293]: DEBUG oslo_concurrency.lockutils [req-a8cf6bda-7c4c-43a3-bfd8-dd44522c85ed req-dcd11307-a83a-4212-9373-7f8e9fef5dfe service nova] Acquiring lock "refresh_cache-fa68f69e-febd-4a46-81b8-afb7e5ce048d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.664132] env[63293]: DEBUG oslo_concurrency.lockutils [req-a8cf6bda-7c4c-43a3-bfd8-dd44522c85ed req-dcd11307-a83a-4212-9373-7f8e9fef5dfe service nova] Acquired lock "refresh_cache-fa68f69e-febd-4a46-81b8-afb7e5ce048d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.664132] env[63293]: DEBUG nova.network.neutron [req-a8cf6bda-7c4c-43a3-bfd8-dd44522c85ed req-dcd11307-a83a-4212-9373-7f8e9fef5dfe service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Refreshing network info cache for port 5e70c362-e59e-422f-b812-98271464078c {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1114.940989] env[63293]: DEBUG oslo_concurrency.lockutils [None req-d63cb30b-b871-4d42-a0ca-3321ad35a6da tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.799s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.941301] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.435s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.941504] env[63293]: INFO nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] During sync_power_state the instance has a pending task (spawning). Skip. [ 1114.941681] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.105219] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.105490] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.107042] env[63293]: INFO nova.compute.claims [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1115.369151] env[63293]: DEBUG nova.network.neutron [req-a8cf6bda-7c4c-43a3-bfd8-dd44522c85ed req-dcd11307-a83a-4212-9373-7f8e9fef5dfe service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Updated VIF entry in instance network info cache for port 5e70c362-e59e-422f-b812-98271464078c. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1115.369529] env[63293]: DEBUG nova.network.neutron [req-a8cf6bda-7c4c-43a3-bfd8-dd44522c85ed req-dcd11307-a83a-4212-9373-7f8e9fef5dfe service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Updating instance_info_cache with network_info: [{"id": "5e70c362-e59e-422f-b812-98271464078c", "address": "fa:16:3e:74:52:dc", "network": {"id": "56e9a0f4-3dbe-4558-a50c-9ca9d32a185b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1632806697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec53642df3804e7190615487a426d4a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e70c362-e5", "ovs_interfaceid": "5e70c362-e59e-422f-b812-98271464078c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.872605] env[63293]: DEBUG oslo_concurrency.lockutils [req-a8cf6bda-7c4c-43a3-bfd8-dd44522c85ed req-dcd11307-a83a-4212-9373-7f8e9fef5dfe service nova] Releasing lock "refresh_cache-fa68f69e-febd-4a46-81b8-afb7e5ce048d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.163347] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b816455e-c6f4-4ac5-8584-a6ea17fee8f8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.170987] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d42908-bf9a-4450-a00c-15aef055b34b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.200976] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ee5b59-a042-43d8-97d4-5a32ad6efe8d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.209077] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9d5954-bedc-42e2-8acd-9abe3883dd2c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.222736] env[63293]: DEBUG nova.compute.provider_tree [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.726347] env[63293]: DEBUG nova.scheduler.client.report [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.232831] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.127s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.233431] env[63293]: DEBUG nova.compute.manager [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Start building networks asynchronously for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1117.739153] env[63293]: DEBUG nova.compute.utils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1117.741506] env[63293]: DEBUG nova.compute.manager [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Allocating IP information in the background. {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1117.741679] env[63293]: DEBUG nova.network.neutron [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] allocate_for_instance() {{(pid=63293) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1117.795480] env[63293]: DEBUG nova.policy [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ebf82e99427d4171a4d510f7d3c966d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1687c33183e74b6ba70e021b6879cb93', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63293) authorize /opt/stack/nova/nova/policy.py:201}} [ 1118.045329] env[63293]: DEBUG nova.network.neutron [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Successfully created port: 10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1118.242321] env[63293]: DEBUG nova.compute.manager [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Start building block device mappings for instance. {{(pid=63293) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1119.252488] env[63293]: DEBUG nova.compute.manager [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Start spawning the instance on the hypervisor. {{(pid=63293) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1119.280029] env[63293]: DEBUG nova.virt.hardware [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T17:31:40Z,direct_url=,disk_format='vmdk',id=1427aa66-93c7-49c6-9e28-dc7fe851dced,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f57886847a1446109c6ab061a248dcf0',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T17:31:41Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1119.280328] env[63293]: DEBUG nova.virt.hardware [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1119.280492] env[63293]: DEBUG nova.virt.hardware [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1119.280680] env[63293]: DEBUG nova.virt.hardware [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1119.280831] env[63293]: DEBUG nova.virt.hardware [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1119.280984] env[63293]: DEBUG nova.virt.hardware [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1119.281216] env[63293]: DEBUG nova.virt.hardware [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1119.281381] env[63293]: DEBUG nova.virt.hardware [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1119.281552] env[63293]: DEBUG nova.virt.hardware [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1119.281717] env[63293]: DEBUG nova.virt.hardware [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1119.281939] env[63293]: DEBUG nova.virt.hardware [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1119.282834] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0dfc795-f8d3-4204-8172-af27d9dfbb48 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.290827] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a270a3-18fa-4842-9e6a-86c65ccdd390 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.427072] env[63293]: DEBUG nova.compute.manager [req-379322ab-6cd6-46cf-998a-09a60bac1268 req-cb2f2913-be11-4cd1-8f3f-09106004908f service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Received event network-vif-plugged-10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1119.427302] env[63293]: DEBUG oslo_concurrency.lockutils [req-379322ab-6cd6-46cf-998a-09a60bac1268 req-cb2f2913-be11-4cd1-8f3f-09106004908f service nova] Acquiring lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.427520] env[63293]: DEBUG oslo_concurrency.lockutils [req-379322ab-6cd6-46cf-998a-09a60bac1268 req-cb2f2913-be11-4cd1-8f3f-09106004908f service nova] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.427698] env[63293]: DEBUG oslo_concurrency.lockutils [req-379322ab-6cd6-46cf-998a-09a60bac1268 req-cb2f2913-be11-4cd1-8f3f-09106004908f service nova] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.427878] env[63293]: DEBUG nova.compute.manager [req-379322ab-6cd6-46cf-998a-09a60bac1268 req-cb2f2913-be11-4cd1-8f3f-09106004908f service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] No waiting events found dispatching network-vif-plugged-10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1119.428109] env[63293]: WARNING nova.compute.manager [req-379322ab-6cd6-46cf-998a-09a60bac1268 req-cb2f2913-be11-4cd1-8f3f-09106004908f service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Received unexpected event network-vif-plugged-10af5b0e-12c0-4492-a508-bd47f02b1173 for instance with vm_state building and task_state spawning. [ 1119.516073] env[63293]: DEBUG nova.network.neutron [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Successfully updated port: 10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1120.019147] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.019302] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.019471] env[63293]: DEBUG nova.network.neutron [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1120.580699] env[63293]: DEBUG nova.network.neutron [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Instance cache missing network info. {{(pid=63293) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1120.702232] env[63293]: DEBUG nova.network.neutron [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updating instance_info_cache with network_info: [{"id": "10af5b0e-12c0-4492-a508-bd47f02b1173", "address": "fa:16:3e:46:6e:33", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10af5b0e-12", "ovs_interfaceid": "10af5b0e-12c0-4492-a508-bd47f02b1173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.205080] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.205293] env[63293]: DEBUG nova.compute.manager [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Instance network_info: |[{"id": "10af5b0e-12c0-4492-a508-bd47f02b1173", "address": "fa:16:3e:46:6e:33", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10af5b0e-12", "ovs_interfaceid": "10af5b0e-12c0-4492-a508-bd47f02b1173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63293) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1121.205742] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:6e:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '10af5b0e-12c0-4492-a508-bd47f02b1173', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1121.213371] env[63293]: DEBUG oslo.service.loopingcall [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1121.213594] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1121.214504] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a69f3311-efa5-4af2-b3bb-c015bc16488e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.234587] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1121.234587] env[63293]: value = "task-1328308" [ 1121.234587] env[63293]: _type = "Task" [ 1121.234587] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.243066] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328308, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.455782] env[63293]: DEBUG nova.compute.manager [req-330c78f5-02c2-4c89-be3f-976532f9e4de req-1ad9303e-7648-4a07-8bcd-cceeb902b201 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Received event network-changed-10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1121.456042] env[63293]: DEBUG nova.compute.manager [req-330c78f5-02c2-4c89-be3f-976532f9e4de req-1ad9303e-7648-4a07-8bcd-cceeb902b201 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Refreshing instance network info cache due to event network-changed-10af5b0e-12c0-4492-a508-bd47f02b1173. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1121.456284] env[63293]: DEBUG oslo_concurrency.lockutils [req-330c78f5-02c2-4c89-be3f-976532f9e4de req-1ad9303e-7648-4a07-8bcd-cceeb902b201 service nova] Acquiring lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.456443] env[63293]: DEBUG oslo_concurrency.lockutils [req-330c78f5-02c2-4c89-be3f-976532f9e4de req-1ad9303e-7648-4a07-8bcd-cceeb902b201 service nova] Acquired lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.456627] env[63293]: DEBUG nova.network.neutron [req-330c78f5-02c2-4c89-be3f-976532f9e4de req-1ad9303e-7648-4a07-8bcd-cceeb902b201 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Refreshing network info cache for port 10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1121.744515] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328308, 'name': CreateVM_Task, 'duration_secs': 0.314381} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.744873] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1121.745418] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.745592] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.745917] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1121.746184] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17d4265b-4d53-4bb9-ab8d-f237ec5854d4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.750621] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1121.750621] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c0da22-8af0-3bf2-e350-0b1c0ab12976" [ 1121.750621] env[63293]: _type = "Task" [ 1121.750621] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.757580] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c0da22-8af0-3bf2-e350-0b1c0ab12976, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.157243] env[63293]: DEBUG nova.network.neutron [req-330c78f5-02c2-4c89-be3f-976532f9e4de req-1ad9303e-7648-4a07-8bcd-cceeb902b201 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updated VIF entry in instance network info cache for port 10af5b0e-12c0-4492-a508-bd47f02b1173. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1122.157616] env[63293]: DEBUG nova.network.neutron [req-330c78f5-02c2-4c89-be3f-976532f9e4de req-1ad9303e-7648-4a07-8bcd-cceeb902b201 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updating instance_info_cache with network_info: [{"id": "10af5b0e-12c0-4492-a508-bd47f02b1173", "address": "fa:16:3e:46:6e:33", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10af5b0e-12", "ovs_interfaceid": "10af5b0e-12c0-4492-a508-bd47f02b1173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.261296] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52c0da22-8af0-3bf2-e350-0b1c0ab12976, 'name': SearchDatastore_Task, 'duration_secs': 0.011157} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.261629] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.261899] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Processing image 1427aa66-93c7-49c6-9e28-dc7fe851dced {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1122.262161] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.262316] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.262499] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1122.262751] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b68edde-e97b-4eee-9559-9eace99f8fe0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.269975] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1122.270169] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1122.270812] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c280559e-68a2-4e2a-b6f3-58d80befdcca {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.275538] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1122.275538] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525eb9e9-f54e-f6b5-f95d-feddd98b5d04" [ 1122.275538] env[63293]: _type = "Task" [ 1122.275538] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.282392] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525eb9e9-f54e-f6b5-f95d-feddd98b5d04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.660997] env[63293]: DEBUG oslo_concurrency.lockutils [req-330c78f5-02c2-4c89-be3f-976532f9e4de req-1ad9303e-7648-4a07-8bcd-cceeb902b201 service nova] Releasing lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.785703] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]525eb9e9-f54e-f6b5-f95d-feddd98b5d04, 'name': SearchDatastore_Task, 'duration_secs': 0.008892} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.786512] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b93b8dd4-3a1a-4793-b72d-606af9fa2372 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.791555] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1122.791555] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a5fcbb-46d8-b2de-03d1-bc1b4478fc6e" [ 1122.791555] env[63293]: _type = "Task" [ 1122.791555] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.798849] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a5fcbb-46d8-b2de-03d1-bc1b4478fc6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.301952] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a5fcbb-46d8-b2de-03d1-bc1b4478fc6e, 'name': SearchDatastore_Task, 'duration_secs': 0.043792} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.302177] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.302443] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 7ae46382-8372-4a33-a9b6-e0720f09811e/7ae46382-8372-4a33-a9b6-e0720f09811e.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1123.302697] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e84dc248-1e27-4cf5-8fa6-e5ff757f2e72 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.309492] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1123.309492] env[63293]: value = "task-1328309" [ 1123.309492] env[63293]: _type = "Task" [ 1123.309492] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.316947] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328309, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.818874] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328309, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492643} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.819285] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1427aa66-93c7-49c6-9e28-dc7fe851dced/1427aa66-93c7-49c6-9e28-dc7fe851dced.vmdk to [datastore2] 7ae46382-8372-4a33-a9b6-e0720f09811e/7ae46382-8372-4a33-a9b6-e0720f09811e.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1123.819431] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Extending root virtual disk to 1048576 {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1123.819694] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71dfb9bc-5176-40d5-9e06-341053f0ba00 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.826527] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1123.826527] env[63293]: value = "task-1328310" [ 1123.826527] env[63293]: _type = "Task" [ 1123.826527] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.833637] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328310, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.336012] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328310, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058469} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.336297] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Extended root virtual disk {{(pid=63293) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1124.337119] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392f67f1-47f1-420c-b1f1-316dcee2123c {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.360175] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 7ae46382-8372-4a33-a9b6-e0720f09811e/7ae46382-8372-4a33-a9b6-e0720f09811e.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.360431] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-679fc040-af25-46e3-b5b6-12a4c981538e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.379479] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1124.379479] env[63293]: value = "task-1328311" [ 1124.379479] env[63293]: _type = "Task" [ 1124.379479] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.386399] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328311, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.889486] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328311, 'name': ReconfigVM_Task, 'duration_secs': 0.2842} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.889878] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 7ae46382-8372-4a33-a9b6-e0720f09811e/7ae46382-8372-4a33-a9b6-e0720f09811e.vmdk or device None with type sparse {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1124.890370] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c0c06335-7068-4b2f-b290-fe40f5217457 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.897524] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1124.897524] env[63293]: value = "task-1328312" [ 1124.897524] env[63293]: _type = "Task" [ 1124.897524] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.904617] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328312, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.407853] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328312, 'name': Rename_Task, 'duration_secs': 0.1389} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.408148] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1125.408394] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d0fa4f8-4c11-455a-bd21-d47743795c05 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.415568] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1125.415568] env[63293]: value = "task-1328313" [ 1125.415568] env[63293]: _type = "Task" [ 1125.415568] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.424724] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328313, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.925219] env[63293]: DEBUG oslo_vmware.api [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328313, 'name': PowerOnVM_Task, 'duration_secs': 0.42788} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.925581] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1125.925705] env[63293]: INFO nova.compute.manager [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Took 6.67 seconds to spawn the instance on the hypervisor. [ 1125.925814] env[63293]: DEBUG nova.compute.manager [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1125.926569] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30c42e4-3626-41c4-9874-66a23411ab15 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.442891] env[63293]: INFO nova.compute.manager [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Took 11.36 seconds to build instance. [ 1126.945406] env[63293]: DEBUG oslo_concurrency.lockutils [None req-bf4ccfac-e9e4-4b0e-9b05-0fefb87b5787 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.865s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.138355] env[63293]: DEBUG nova.compute.manager [req-e83ea560-19aa-49ae-9352-15cb2569386c req-404b5847-d1fd-4195-9bd4-f7aeb80a2806 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Received event network-changed-10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1127.138512] env[63293]: DEBUG nova.compute.manager [req-e83ea560-19aa-49ae-9352-15cb2569386c req-404b5847-d1fd-4195-9bd4-f7aeb80a2806 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Refreshing instance network info cache due to event network-changed-10af5b0e-12c0-4492-a508-bd47f02b1173. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1127.138735] env[63293]: DEBUG oslo_concurrency.lockutils [req-e83ea560-19aa-49ae-9352-15cb2569386c req-404b5847-d1fd-4195-9bd4-f7aeb80a2806 service nova] Acquiring lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.138884] env[63293]: DEBUG oslo_concurrency.lockutils [req-e83ea560-19aa-49ae-9352-15cb2569386c req-404b5847-d1fd-4195-9bd4-f7aeb80a2806 service nova] Acquired lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.139063] env[63293]: DEBUG nova.network.neutron [req-e83ea560-19aa-49ae-9352-15cb2569386c req-404b5847-d1fd-4195-9bd4-f7aeb80a2806 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Refreshing network info cache for port 10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1127.841990] env[63293]: DEBUG nova.network.neutron [req-e83ea560-19aa-49ae-9352-15cb2569386c req-404b5847-d1fd-4195-9bd4-f7aeb80a2806 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updated VIF entry in instance network info cache for port 10af5b0e-12c0-4492-a508-bd47f02b1173. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1127.842416] env[63293]: DEBUG nova.network.neutron [req-e83ea560-19aa-49ae-9352-15cb2569386c req-404b5847-d1fd-4195-9bd4-f7aeb80a2806 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updating instance_info_cache with network_info: [{"id": "10af5b0e-12c0-4492-a508-bd47f02b1173", "address": "fa:16:3e:46:6e:33", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10af5b0e-12", "ovs_interfaceid": "10af5b0e-12c0-4492-a508-bd47f02b1173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.345173] env[63293]: DEBUG oslo_concurrency.lockutils [req-e83ea560-19aa-49ae-9352-15cb2569386c req-404b5847-d1fd-4195-9bd4-f7aeb80a2806 service nova] Releasing lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.870958] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.871353] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63293) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1130.866880] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1130.871571] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1131.870972] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.373722] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.373999] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.374177] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.374355] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63293) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1132.375266] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de142643-3bf3-4439-9fa2-c669c711cec7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.384433] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90437890-fcbe-447a-ba7a-768b913e5bbf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.397809] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b2c9d9-7df9-4847-aa9a-9159d946451b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.403661] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66260b98-0260-4655-a78a-92ce21b57267 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.430941] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181073MB free_disk=132GB free_vcpus=48 pci_devices=None {{(pid=63293) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1132.431145] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.431265] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.456557] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1133.456857] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance fa68f69e-febd-4a46-81b8-afb7e5ce048d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1133.456857] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 7ae46382-8372-4a33-a9b6-e0720f09811e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1133.457017] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1133.457197] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1133.505215] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77fb0c87-6e6e-4b6d-ae73-212569b5f642 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.512767] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84863580-f0b6-40b7-b9cb-da2424d805b2 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.543840] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4626dac5-8304-4121-840a-d0473a23facf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.550807] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d5ef3d-863e-46b3-887b-37fd4ce23cde {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.563507] env[63293]: DEBUG nova.compute.provider_tree [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.066244] env[63293]: DEBUG nova.scheduler.client.report [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1134.571063] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1134.571479] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.140s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.571727] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.572164] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Starting heal instance info cache {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1136.102153] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.102323] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquired lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.102470] env[63293]: DEBUG nova.network.neutron [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Forcefully refreshing network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1137.326198] env[63293]: DEBUG nova.network.neutron [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updating instance_info_cache with network_info: [{"id": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "address": "fa:16:3e:85:83:c3", "network": {"id": "bf5244bc-1a96-4460-ae51-d8731c202a8f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-13669700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "affba61d6a2846b38666544bc2c25db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24144f5a-050a-4f1e-8d8c-774dc16dc791", "external-id": "cl2-zone-252", "segmentation_id": 252, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3313c3a6-17", "ovs_interfaceid": "3313c3a6-17f4-4c97-b19f-9edfd6eb4099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.828587] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Releasing lock "refresh_cache-97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.828775] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updated the network info_cache for instance {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1137.828965] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.829147] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.829298] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.829447] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.650915] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.651324] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.651460] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.651626] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.651796] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.655224] env[63293]: INFO nova.compute.manager [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Terminating instance [ 1144.656949] env[63293]: DEBUG nova.compute.manager [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1144.657160] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1144.657980] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50040008-e373-40f4-9ed4-def5c4d87bd4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.665958] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1144.666192] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-796ef29b-5fb7-447a-8507-15f52ab2c9b4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.672192] env[63293]: DEBUG oslo_vmware.api [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1144.672192] env[63293]: value = "task-1328314" [ 1144.672192] env[63293]: _type = "Task" [ 1144.672192] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.679448] env[63293]: DEBUG oslo_vmware.api [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328314, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.182258] env[63293]: DEBUG oslo_vmware.api [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328314, 'name': PowerOffVM_Task, 'duration_secs': 0.189482} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.182538] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1145.182714] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1145.182960] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1543fb3-3c7c-4191-a172-3238442eab20 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.245121] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1145.245344] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Deleting contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1145.245529] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleting the datastore file [datastore1] 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1145.245795] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fda2412-ba4a-48fe-8bdd-eb266914d7fa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.252847] env[63293]: DEBUG oslo_vmware.api [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for the task: (returnval){ [ 1145.252847] env[63293]: value = "task-1328316" [ 1145.252847] env[63293]: _type = "Task" [ 1145.252847] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.259932] env[63293]: DEBUG oslo_vmware.api [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.762570] env[63293]: DEBUG oslo_vmware.api [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Task: {'id': task-1328316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13191} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.762973] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1145.763026] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Deleted contents of the VM from datastore datastore1 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1145.763198] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1145.763376] env[63293]: INFO nova.compute.manager [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1145.763624] env[63293]: DEBUG oslo.service.loopingcall [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1145.763814] env[63293]: DEBUG nova.compute.manager [-] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1145.763907] env[63293]: DEBUG nova.network.neutron [-] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1146.265244] env[63293]: DEBUG nova.compute.manager [req-4d1a5914-4203-404f-948f-7ec7571142a1 req-5e9705ce-ad80-4b59-81e7-2198d9bc72ce service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Received event network-vif-deleted-3313c3a6-17f4-4c97-b19f-9edfd6eb4099 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1146.265451] env[63293]: INFO nova.compute.manager [req-4d1a5914-4203-404f-948f-7ec7571142a1 req-5e9705ce-ad80-4b59-81e7-2198d9bc72ce service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Neutron deleted interface 3313c3a6-17f4-4c97-b19f-9edfd6eb4099; detaching it from the instance and deleting it from the info cache [ 1146.265649] env[63293]: DEBUG nova.network.neutron [req-4d1a5914-4203-404f-948f-7ec7571142a1 req-5e9705ce-ad80-4b59-81e7-2198d9bc72ce service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.717699] env[63293]: DEBUG nova.network.neutron [-] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.770030] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a6e0b765-b366-46b7-8e98-0c4b63fd96fe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.780345] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d1bba93-a7d4-46ce-92b5-bf4a5e3954cb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.803796] env[63293]: DEBUG nova.compute.manager [req-4d1a5914-4203-404f-948f-7ec7571142a1 req-5e9705ce-ad80-4b59-81e7-2198d9bc72ce service nova] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Detach interface failed, port_id=3313c3a6-17f4-4c97-b19f-9edfd6eb4099, reason: Instance 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1147.221131] env[63293]: INFO nova.compute.manager [-] [instance: 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d] Took 1.46 seconds to deallocate network for instance. [ 1147.727216] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.727472] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.727702] env[63293]: DEBUG nova.objects.instance [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lazy-loading 'resources' on Instance uuid 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1148.281546] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775ecc31-83fe-4cc7-b247-8194d3c3a4d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.288852] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0766c31d-75d1-4fad-bbd7-979deef93bcd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.317921] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226dd338-9dc0-4f0b-98e3-373af5df592d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.324423] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0232e46a-2c75-419b-b2a6-d672fdfb7636 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.337183] env[63293]: DEBUG nova.compute.provider_tree [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.840484] env[63293]: DEBUG nova.scheduler.client.report [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1149.345658] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.367402] env[63293]: INFO nova.scheduler.client.report [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Deleted allocations for instance 97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d [ 1149.876054] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ff8c4835-35a6-4f4f-944a-ac646a08ca51 tempest-AttachVolumeShelveTestJSON-695140506 tempest-AttachVolumeShelveTestJSON-695140506-project-member] Lock "97d0fb6c-2dc9-44ae-96f3-01e3f4c9f03d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.224s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.256783] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.257110] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.760892] env[63293]: DEBUG nova.compute.utils [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Using /dev/sd instead of None {{(pid=63293) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1154.263853] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.326326] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.326627] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.326854] env[63293]: INFO nova.compute.manager [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Attaching volume 4963d5ee-41e7-4986-bed7-73b8a45187ee to /dev/sdb [ 1155.357302] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4132eb1c-752b-474f-9905-063be59aab1d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.364377] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31248a1e-adb8-4085-bb81-f01bc44e1f2a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.376900] env[63293]: DEBUG nova.virt.block_device [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Updating existing volume attachment record: bb0b8f64-4a58-4c09-9bd6-400396ffab02 {{(pid=63293) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1159.919618] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Volume attach. Driver type: vmdk {{(pid=63293) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1159.920115] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283857', 'volume_id': '4963d5ee-41e7-4986-bed7-73b8a45187ee', 'name': 'volume-4963d5ee-41e7-4986-bed7-73b8a45187ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fa68f69e-febd-4a46-81b8-afb7e5ce048d', 'attached_at': '', 'detached_at': '', 'volume_id': '4963d5ee-41e7-4986-bed7-73b8a45187ee', 'serial': '4963d5ee-41e7-4986-bed7-73b8a45187ee'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1159.921253] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8d265e-782a-4b49-a052-0c2d682253ac {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.938923] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d23278-af91-4aab-b80e-5f732114645d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.962723] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] volume-4963d5ee-41e7-4986-bed7-73b8a45187ee/volume-4963d5ee-41e7-4986-bed7-73b8a45187ee.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1159.962976] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdf3dd11-d2a3-44d1-9d7d-0bc5a6a161a6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.980887] env[63293]: DEBUG oslo_vmware.api [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1159.980887] env[63293]: value = "task-1328320" [ 1159.980887] env[63293]: _type = "Task" [ 1159.980887] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.988514] env[63293]: DEBUG oslo_vmware.api [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328320, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.491234] env[63293]: DEBUG oslo_vmware.api [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328320, 'name': ReconfigVM_Task, 'duration_secs': 0.34744} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.491526] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Reconfigured VM instance instance-0000006d to attach disk [datastore1] volume-4963d5ee-41e7-4986-bed7-73b8a45187ee/volume-4963d5ee-41e7-4986-bed7-73b8a45187ee.vmdk or device None with type thin {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1160.496111] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6652203-bbf2-4c41-95d3-8665783c85e6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.511019] env[63293]: DEBUG oslo_vmware.api [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1160.511019] env[63293]: value = "task-1328321" [ 1160.511019] env[63293]: _type = "Task" [ 1160.511019] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.519572] env[63293]: DEBUG oslo_vmware.api [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328321, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.025186] env[63293]: DEBUG oslo_vmware.api [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328321, 'name': ReconfigVM_Task, 'duration_secs': 0.127115} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.025685] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283857', 'volume_id': '4963d5ee-41e7-4986-bed7-73b8a45187ee', 'name': 'volume-4963d5ee-41e7-4986-bed7-73b8a45187ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fa68f69e-febd-4a46-81b8-afb7e5ce048d', 'attached_at': '', 'detached_at': '', 'volume_id': '4963d5ee-41e7-4986-bed7-73b8a45187ee', 'serial': '4963d5ee-41e7-4986-bed7-73b8a45187ee'} {{(pid=63293) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1162.061811] env[63293]: DEBUG nova.objects.instance [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lazy-loading 'flavor' on Instance uuid fa68f69e-febd-4a46-81b8-afb7e5ce048d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1162.567706] env[63293]: DEBUG oslo_concurrency.lockutils [None req-5bb6e478-e56c-4d28-b66c-20b103e316ad tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.241s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.744975] env[63293]: DEBUG oslo_concurrency.lockutils [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.745267] env[63293]: DEBUG oslo_concurrency.lockutils [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.248947] env[63293]: INFO nova.compute.manager [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Detaching volume 4963d5ee-41e7-4986-bed7-73b8a45187ee [ 1163.277437] env[63293]: INFO nova.virt.block_device [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Attempting to driver detach volume 4963d5ee-41e7-4986-bed7-73b8a45187ee from mountpoint /dev/sdb [ 1163.277437] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Volume detach. Driver type: vmdk {{(pid=63293) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1163.277600] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283857', 'volume_id': '4963d5ee-41e7-4986-bed7-73b8a45187ee', 'name': 'volume-4963d5ee-41e7-4986-bed7-73b8a45187ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fa68f69e-febd-4a46-81b8-afb7e5ce048d', 'attached_at': '', 'detached_at': '', 'volume_id': '4963d5ee-41e7-4986-bed7-73b8a45187ee', 'serial': '4963d5ee-41e7-4986-bed7-73b8a45187ee'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1163.278494] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513f58d0-4462-4bdc-b336-9a3bc9d3d614 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.300638] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec89f7b5-c877-4fc0-af40-192013e8335e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.307244] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe2d877-91c5-4059-b0de-17bb04e56ede {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.326661] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ecb660d-5e9d-4d2d-a7cd-a28900f5dd08 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.342290] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] The volume has not been displaced from its original location: [datastore1] volume-4963d5ee-41e7-4986-bed7-73b8a45187ee/volume-4963d5ee-41e7-4986-bed7-73b8a45187ee.vmdk. No consolidation needed. {{(pid=63293) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1163.347520] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1163.348793] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b68de9e-cdaf-4f57-9550-6962d73d4a8a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.368807] env[63293]: DEBUG oslo_vmware.api [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1163.368807] env[63293]: value = "task-1328322" [ 1163.368807] env[63293]: _type = "Task" [ 1163.368807] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.378507] env[63293]: DEBUG oslo_vmware.api [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328322, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.865370] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "7ae46382-8372-4a33-a9b6-e0720f09811e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.865710] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.865881] env[63293]: INFO nova.compute.manager [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Shelving [ 1163.879367] env[63293]: DEBUG oslo_vmware.api [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328322, 'name': ReconfigVM_Task, 'duration_secs': 0.203258} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.879632] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=63293) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1163.884671] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd1953b9-f00c-4b02-8a51-84b99972f4a5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.900070] env[63293]: DEBUG oslo_vmware.api [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1163.900070] env[63293]: value = "task-1328323" [ 1163.900070] env[63293]: _type = "Task" [ 1163.900070] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.907696] env[63293]: DEBUG oslo_vmware.api [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328323, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.373218] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1164.373540] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2e5948b-926b-4f39-9cfb-aa8960167d65 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.382180] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1164.382180] env[63293]: value = "task-1328324" [ 1164.382180] env[63293]: _type = "Task" [ 1164.382180] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.390296] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328324, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.408907] env[63293]: DEBUG oslo_vmware.api [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328323, 'name': ReconfigVM_Task, 'duration_secs': 0.126394} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.409281] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-283857', 'volume_id': '4963d5ee-41e7-4986-bed7-73b8a45187ee', 'name': 'volume-4963d5ee-41e7-4986-bed7-73b8a45187ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fa68f69e-febd-4a46-81b8-afb7e5ce048d', 'attached_at': '', 'detached_at': '', 'volume_id': '4963d5ee-41e7-4986-bed7-73b8a45187ee', 'serial': '4963d5ee-41e7-4986-bed7-73b8a45187ee'} {{(pid=63293) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1164.892304] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328324, 'name': PowerOffVM_Task, 'duration_secs': 0.190179} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.892575] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1164.893364] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-479bd26f-30ea-4a44-8e68-da12cacd7054 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.910905] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6526d75f-18e1-447e-a8b7-4aa5f679aa91 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.954877] env[63293]: DEBUG nova.objects.instance [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lazy-loading 'flavor' on Instance uuid fa68f69e-febd-4a46-81b8-afb7e5ce048d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.421503] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Creating Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1165.422219] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-da07c8d1-65bd-48ee-9f44-0faa396b3605 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.430741] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1165.430741] env[63293]: value = "task-1328325" [ 1165.430741] env[63293]: _type = "Task" [ 1165.430741] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.438938] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328325, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.940548] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328325, 'name': CreateSnapshot_Task, 'duration_secs': 0.402696} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.940781] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Created Snapshot of the VM instance {{(pid=63293) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1165.941542] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d98d856-ec7d-4928-b2cf-07a821e31cd4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.962449] env[63293]: DEBUG oslo_concurrency.lockutils [None req-22de3fbf-8aa0-42bc-9a45-1e67d09cfe8a tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.217s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.458434] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Creating linked-clone VM from snapshot {{(pid=63293) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1166.458783] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cfe307e1-996a-4f9b-bf7b-5c3eb97c1e24 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.469206] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1166.469206] env[63293]: value = "task-1328326" [ 1166.469206] env[63293]: _type = "Task" [ 1166.469206] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.477279] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328326, 'name': CloneVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.979546] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328326, 'name': CloneVM_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.986113] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.986352] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.986602] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.986818] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.987057] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.989382] env[63293]: INFO nova.compute.manager [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Terminating instance [ 1166.991245] env[63293]: DEBUG nova.compute.manager [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1166.991445] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1166.992211] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d977b594-60a4-435e-a93c-39929e1c14e6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.999178] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1166.999405] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6164f2c5-9542-4b09-8394-d7c528d60f98 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.005343] env[63293]: DEBUG oslo_vmware.api [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1167.005343] env[63293]: value = "task-1328327" [ 1167.005343] env[63293]: _type = "Task" [ 1167.005343] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.012939] env[63293]: DEBUG oslo_vmware.api [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328327, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.479724] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328326, 'name': CloneVM_Task, 'duration_secs': 0.918396} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.480153] env[63293]: INFO nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Created linked-clone VM from snapshot [ 1167.480741] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89abe051-bfd9-4a3d-8ee2-c79f97b15ac5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.487596] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Uploading image 5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1167.512847] env[63293]: DEBUG oslo_vmware.rw_handles [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1167.512847] env[63293]: value = "vm-283859" [ 1167.512847] env[63293]: _type = "VirtualMachine" [ 1167.512847] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1167.513105] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5a8345b9-1638-41dc-b6b4-33acc44ae1cf {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.517518] env[63293]: DEBUG oslo_vmware.api [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328327, 'name': PowerOffVM_Task, 'duration_secs': 0.177774} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.518100] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1167.518290] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1167.518519] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6bbd540e-40df-4c47-b62f-464528ca50bd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.522759] env[63293]: DEBUG oslo_vmware.rw_handles [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lease: (returnval){ [ 1167.522759] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52aeeb80-f4a1-e4f7-77bc-f8418319d2b7" [ 1167.522759] env[63293]: _type = "HttpNfcLease" [ 1167.522759] env[63293]: } obtained for exporting VM: (result){ [ 1167.522759] env[63293]: value = "vm-283859" [ 1167.522759] env[63293]: _type = "VirtualMachine" [ 1167.522759] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1167.523015] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the lease: (returnval){ [ 1167.523015] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52aeeb80-f4a1-e4f7-77bc-f8418319d2b7" [ 1167.523015] env[63293]: _type = "HttpNfcLease" [ 1167.523015] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1167.528563] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1167.528563] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52aeeb80-f4a1-e4f7-77bc-f8418319d2b7" [ 1167.528563] env[63293]: _type = "HttpNfcLease" [ 1167.528563] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1167.662016] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1167.662347] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1167.662621] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleting the datastore file [datastore2] fa68f69e-febd-4a46-81b8-afb7e5ce048d {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1167.662940] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d41c58c9-0d62-401a-bc9a-b1a2e10d6e6b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.670778] env[63293]: DEBUG oslo_vmware.api [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for the task: (returnval){ [ 1167.670778] env[63293]: value = "task-1328330" [ 1167.670778] env[63293]: _type = "Task" [ 1167.670778] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.679880] env[63293]: DEBUG oslo_vmware.api [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.031590] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1168.031590] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52aeeb80-f4a1-e4f7-77bc-f8418319d2b7" [ 1168.031590] env[63293]: _type = "HttpNfcLease" [ 1168.031590] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1168.031892] env[63293]: DEBUG oslo_vmware.rw_handles [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1168.031892] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52aeeb80-f4a1-e4f7-77bc-f8418319d2b7" [ 1168.031892] env[63293]: _type = "HttpNfcLease" [ 1168.031892] env[63293]: }. {{(pid=63293) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1168.032632] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58576a58-499d-4b4f-9583-4832ab2a6653 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.039740] env[63293]: DEBUG oslo_vmware.rw_handles [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526af53a-caf4-69da-fda6-5f0adb216c32/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1168.039917] env[63293]: DEBUG oslo_vmware.rw_handles [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526af53a-caf4-69da-fda6-5f0adb216c32/disk-0.vmdk for reading. {{(pid=63293) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1168.130397] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d5d474dd-7932-4070-86f3-ef07f7f0ef55 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.180097] env[63293]: DEBUG oslo_vmware.api [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Task: {'id': task-1328330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134159} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.180368] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1168.180567] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1168.180743] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1168.180918] env[63293]: INFO nova.compute.manager [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1168.181201] env[63293]: DEBUG oslo.service.loopingcall [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1168.181400] env[63293]: DEBUG nova.compute.manager [-] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1168.181503] env[63293]: DEBUG nova.network.neutron [-] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1168.659765] env[63293]: DEBUG nova.compute.manager [req-41cf39b1-3ab7-4660-aed6-51d831c108d9 req-03fdbfd7-1685-474f-b6f8-344799332a06 service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Received event network-vif-deleted-5e70c362-e59e-422f-b812-98271464078c {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1168.660969] env[63293]: INFO nova.compute.manager [req-41cf39b1-3ab7-4660-aed6-51d831c108d9 req-03fdbfd7-1685-474f-b6f8-344799332a06 service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Neutron deleted interface 5e70c362-e59e-422f-b812-98271464078c; detaching it from the instance and deleting it from the info cache [ 1168.660969] env[63293]: DEBUG nova.network.neutron [req-41cf39b1-3ab7-4660-aed6-51d831c108d9 req-03fdbfd7-1685-474f-b6f8-344799332a06 service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.134582] env[63293]: DEBUG nova.network.neutron [-] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.165773] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-247d8db4-df19-4ae5-9c80-83df060606f0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.176220] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbaa6db-3e67-467b-a853-56569b998480 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.200976] env[63293]: DEBUG nova.compute.manager [req-41cf39b1-3ab7-4660-aed6-51d831c108d9 req-03fdbfd7-1685-474f-b6f8-344799332a06 service nova] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Detach interface failed, port_id=5e70c362-e59e-422f-b812-98271464078c, reason: Instance fa68f69e-febd-4a46-81b8-afb7e5ce048d could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1169.638107] env[63293]: INFO nova.compute.manager [-] [instance: fa68f69e-febd-4a46-81b8-afb7e5ce048d] Took 1.46 seconds to deallocate network for instance. [ 1170.145650] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.146983] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.146983] env[63293]: DEBUG nova.objects.instance [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lazy-loading 'resources' on Instance uuid fa68f69e-febd-4a46-81b8-afb7e5ce048d {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1170.695465] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5a6720-1875-4f20-9bf6-4793d36a2a30 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.703648] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ac0c35-c888-4f6a-a45c-5a0ff8082e1f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.736474] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b4f361b-a517-4850-be39-d0845b5ce2ec {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.743596] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2d85bd-a6f0-456f-8e5f-816b187934d0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.756287] env[63293]: DEBUG nova.compute.provider_tree [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1171.259861] env[63293]: DEBUG nova.scheduler.client.report [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1171.765884] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.788855] env[63293]: INFO nova.scheduler.client.report [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Deleted allocations for instance fa68f69e-febd-4a46-81b8-afb7e5ce048d [ 1172.298275] env[63293]: DEBUG oslo_concurrency.lockutils [None req-ca524afa-9b87-4f5a-aa6a-c70a054b3975 tempest-AttachVolumeNegativeTest-136651543 tempest-AttachVolumeNegativeTest-136651543-project-member] Lock "fa68f69e-febd-4a46-81b8-afb7e5ce048d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.312s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.207504] env[63293]: DEBUG oslo_vmware.rw_handles [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526af53a-caf4-69da-fda6-5f0adb216c32/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1175.208446] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babb7c49-f271-43ae-a7be-6a0cdbd03a29 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.214909] env[63293]: DEBUG oslo_vmware.rw_handles [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526af53a-caf4-69da-fda6-5f0adb216c32/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1175.215087] env[63293]: ERROR oslo_vmware.rw_handles [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526af53a-caf4-69da-fda6-5f0adb216c32/disk-0.vmdk due to incomplete transfer. [ 1175.215313] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-82d0a2f5-8f7f-4a80-bc6f-1dfb5b9186d9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.221951] env[63293]: DEBUG oslo_vmware.rw_handles [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526af53a-caf4-69da-fda6-5f0adb216c32/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1175.222166] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Uploaded image 5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f to the Glance image server {{(pid=63293) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1175.224401] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Destroying the VM {{(pid=63293) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1175.224632] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a60cf394-e088-4f1a-9278-e81e5cb92626 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.231048] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1175.231048] env[63293]: value = "task-1328333" [ 1175.231048] env[63293]: _type = "Task" [ 1175.231048] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.238591] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328333, 'name': Destroy_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.741802] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328333, 'name': Destroy_Task, 'duration_secs': 0.30359} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.742163] env[63293]: INFO nova.virt.vmwareapi.vm_util [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Destroyed the VM [ 1175.742415] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Deleting Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1175.742703] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c00e63c7-2d39-4b78-8763-86885d9b956d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.748581] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1175.748581] env[63293]: value = "task-1328334" [ 1175.748581] env[63293]: _type = "Task" [ 1175.748581] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.756170] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328334, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.259894] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328334, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.758947] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328334, 'name': RemoveSnapshot_Task, 'duration_secs': 0.602235} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.759251] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Deleted Snapshot of the VM instance {{(pid=63293) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1176.759532] env[63293]: DEBUG nova.compute.manager [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1176.760303] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b8c1c4-d804-4d9b-a84b-006da0aaaa8a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.271841] env[63293]: INFO nova.compute.manager [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Shelve offloading [ 1177.273591] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1177.273823] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-adcb3c17-ebd1-4fdd-861d-164574f738c7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.280934] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1177.280934] env[63293]: value = "task-1328335" [ 1177.280934] env[63293]: _type = "Task" [ 1177.280934] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.288155] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328335, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.792171] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] VM already powered off {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1177.792950] env[63293]: DEBUG nova.compute.manager [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1177.793213] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb44a9d-6e82-4cd4-94ea-e6edbbeb93ba {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.798930] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1177.799104] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.799271] env[63293]: DEBUG nova.network.neutron [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1178.518051] env[63293]: DEBUG nova.network.neutron [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updating instance_info_cache with network_info: [{"id": "10af5b0e-12c0-4492-a508-bd47f02b1173", "address": "fa:16:3e:46:6e:33", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10af5b0e-12", "ovs_interfaceid": "10af5b0e-12c0-4492-a508-bd47f02b1173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.021397] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.312769] env[63293]: DEBUG nova.compute.manager [req-7885ea70-f42c-4f1b-ba9f-a807ef390c94 req-6c9ac362-5673-433f-bced-1b19d158fa52 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Received event network-vif-unplugged-10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1179.313124] env[63293]: DEBUG oslo_concurrency.lockutils [req-7885ea70-f42c-4f1b-ba9f-a807ef390c94 req-6c9ac362-5673-433f-bced-1b19d158fa52 service nova] Acquiring lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.313432] env[63293]: DEBUG oslo_concurrency.lockutils [req-7885ea70-f42c-4f1b-ba9f-a807ef390c94 req-6c9ac362-5673-433f-bced-1b19d158fa52 service nova] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.313698] env[63293]: DEBUG oslo_concurrency.lockutils [req-7885ea70-f42c-4f1b-ba9f-a807ef390c94 req-6c9ac362-5673-433f-bced-1b19d158fa52 service nova] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.314029] env[63293]: DEBUG nova.compute.manager [req-7885ea70-f42c-4f1b-ba9f-a807ef390c94 req-6c9ac362-5673-433f-bced-1b19d158fa52 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] No waiting events found dispatching network-vif-unplugged-10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1179.314591] env[63293]: WARNING nova.compute.manager [req-7885ea70-f42c-4f1b-ba9f-a807ef390c94 req-6c9ac362-5673-433f-bced-1b19d158fa52 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Received unexpected event network-vif-unplugged-10af5b0e-12c0-4492-a508-bd47f02b1173 for instance with vm_state shelved and task_state shelving_offloading. [ 1179.398360] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1179.399399] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367a9763-d5e2-47ac-8b9a-e59495d047e7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.407171] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1179.407459] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2bc70da2-70fc-4b9a-9280-cc2448c64aa3 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.485570] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1179.485837] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1179.486040] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleting the datastore file [datastore2] 7ae46382-8372-4a33-a9b6-e0720f09811e {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1179.486333] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7dc76620-cdff-40df-b587-67e66cbcf90f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.493209] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1179.493209] env[63293]: value = "task-1328338" [ 1179.493209] env[63293]: _type = "Task" [ 1179.493209] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.501139] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.003061] env[63293]: DEBUG oslo_vmware.api [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15296} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.003386] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1180.003538] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1180.003752] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1180.023923] env[63293]: INFO nova.scheduler.client.report [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleted allocations for instance 7ae46382-8372-4a33-a9b6-e0720f09811e [ 1180.528722] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1180.529061] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.529262] env[63293]: DEBUG nova.objects.instance [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'resources' on Instance uuid 7ae46382-8372-4a33-a9b6-e0720f09811e {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1181.033665] env[63293]: DEBUG nova.objects.instance [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'numa_topology' on Instance uuid 7ae46382-8372-4a33-a9b6-e0720f09811e {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1181.345299] env[63293]: DEBUG nova.compute.manager [req-22419de7-b660-4888-b1b4-76c793ed02e9 req-b5a23762-f9ef-46b5-a6f9-b363776ab29e service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Received event network-changed-10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1181.345611] env[63293]: DEBUG nova.compute.manager [req-22419de7-b660-4888-b1b4-76c793ed02e9 req-b5a23762-f9ef-46b5-a6f9-b363776ab29e service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Refreshing instance network info cache due to event network-changed-10af5b0e-12c0-4492-a508-bd47f02b1173. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1181.345611] env[63293]: DEBUG oslo_concurrency.lockutils [req-22419de7-b660-4888-b1b4-76c793ed02e9 req-b5a23762-f9ef-46b5-a6f9-b363776ab29e service nova] Acquiring lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.345745] env[63293]: DEBUG oslo_concurrency.lockutils [req-22419de7-b660-4888-b1b4-76c793ed02e9 req-b5a23762-f9ef-46b5-a6f9-b363776ab29e service nova] Acquired lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.345992] env[63293]: DEBUG nova.network.neutron [req-22419de7-b660-4888-b1b4-76c793ed02e9 req-b5a23762-f9ef-46b5-a6f9-b363776ab29e service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Refreshing network info cache for port 10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1181.535986] env[63293]: DEBUG nova.objects.base [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Object Instance<7ae46382-8372-4a33-a9b6-e0720f09811e> lazy-loaded attributes: resources,numa_topology {{(pid=63293) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1181.568043] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b5bf89-e738-42d7-ab94-8ab624fa6853 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.575868] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b98a702-15b2-4fba-82c5-4faebe77b6ad {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.605862] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babe1da0-ecdb-40ee-9032-9dbe523ce79d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.613997] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73152a73-50d5-49f2-80f0-9132a305bcd4 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.627797] env[63293]: DEBUG nova.compute.provider_tree [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1181.684690] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "7ae46382-8372-4a33-a9b6-e0720f09811e" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.070853] env[63293]: DEBUG nova.network.neutron [req-22419de7-b660-4888-b1b4-76c793ed02e9 req-b5a23762-f9ef-46b5-a6f9-b363776ab29e service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updated VIF entry in instance network info cache for port 10af5b0e-12c0-4492-a508-bd47f02b1173. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1182.071244] env[63293]: DEBUG nova.network.neutron [req-22419de7-b660-4888-b1b4-76c793ed02e9 req-b5a23762-f9ef-46b5-a6f9-b363776ab29e service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updating instance_info_cache with network_info: [{"id": "10af5b0e-12c0-4492-a508-bd47f02b1173", "address": "fa:16:3e:46:6e:33", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap10af5b0e-12", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.130756] env[63293]: DEBUG nova.scheduler.client.report [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1182.574399] env[63293]: DEBUG oslo_concurrency.lockutils [req-22419de7-b660-4888-b1b4-76c793ed02e9 req-b5a23762-f9ef-46b5-a6f9-b363776ab29e service nova] Releasing lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.635583] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.106s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.143570] env[63293]: DEBUG oslo_concurrency.lockutils [None req-53937de5-20fe-414c-a459-947a536c2928 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.278s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.144359] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.460s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.144550] env[63293]: INFO nova.compute.manager [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Unshelving [ 1184.170394] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.170723] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.170958] env[63293]: DEBUG nova.objects.instance [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'pci_requests' on Instance uuid 7ae46382-8372-4a33-a9b6-e0720f09811e {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1184.675194] env[63293]: DEBUG nova.objects.instance [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'numa_topology' on Instance uuid 7ae46382-8372-4a33-a9b6-e0720f09811e {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1185.178399] env[63293]: INFO nova.compute.claims [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1186.213723] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26463917-8906-4b78-b55d-374bf192cbb6 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.221155] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d000f421-5cc4-4fe4-9c22-37f02fc8d9a9 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.252955] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a559145-9423-462b-a211-ae9632f92b47 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.259833] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3792c723-ebcc-45d8-8be4-fd224321f904 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.272495] env[63293]: DEBUG nova.compute.provider_tree [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1186.775326] env[63293]: DEBUG nova.scheduler.client.report [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1187.279936] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.109s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.323531] env[63293]: INFO nova.network.neutron [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updating port 10af5b0e-12c0-4492-a508-bd47f02b1173 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1188.698144] env[63293]: DEBUG nova.compute.manager [req-e5e0d2a3-89e9-4063-bbed-48699526a365 req-73f0690e-dce5-49b0-bdaf-bb57c4840d48 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Received event network-vif-plugged-10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1188.698144] env[63293]: DEBUG oslo_concurrency.lockutils [req-e5e0d2a3-89e9-4063-bbed-48699526a365 req-73f0690e-dce5-49b0-bdaf-bb57c4840d48 service nova] Acquiring lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.698144] env[63293]: DEBUG oslo_concurrency.lockutils [req-e5e0d2a3-89e9-4063-bbed-48699526a365 req-73f0690e-dce5-49b0-bdaf-bb57c4840d48 service nova] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.698144] env[63293]: DEBUG oslo_concurrency.lockutils [req-e5e0d2a3-89e9-4063-bbed-48699526a365 req-73f0690e-dce5-49b0-bdaf-bb57c4840d48 service nova] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.698935] env[63293]: DEBUG nova.compute.manager [req-e5e0d2a3-89e9-4063-bbed-48699526a365 req-73f0690e-dce5-49b0-bdaf-bb57c4840d48 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] No waiting events found dispatching network-vif-plugged-10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1188.699308] env[63293]: WARNING nova.compute.manager [req-e5e0d2a3-89e9-4063-bbed-48699526a365 req-73f0690e-dce5-49b0-bdaf-bb57c4840d48 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Received unexpected event network-vif-plugged-10af5b0e-12c0-4492-a508-bd47f02b1173 for instance with vm_state shelved_offloaded and task_state spawning. [ 1188.777946] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1188.777946] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1188.777946] env[63293]: DEBUG nova.network.neutron [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Building network info cache for instance {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1189.475785] env[63293]: DEBUG nova.network.neutron [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updating instance_info_cache with network_info: [{"id": "10af5b0e-12c0-4492-a508-bd47f02b1173", "address": "fa:16:3e:46:6e:33", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10af5b0e-12", "ovs_interfaceid": "10af5b0e-12c0-4492-a508-bd47f02b1173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.871684] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.872048] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63293) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1189.978858] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.007819] env[63293]: DEBUG nova.virt.hardware [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T17:31:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='8494e3c0c842c1e22e0688125c9017ea',container_format='bare',created_at=2024-10-21T17:44:26Z,direct_url=,disk_format='vmdk',id=5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-837988255-shelved',owner='1687c33183e74b6ba70e021b6879cb93',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2024-10-21T17:44:38Z,virtual_size=,visibility=), allow threads: False {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1190.008121] env[63293]: DEBUG nova.virt.hardware [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1190.008294] env[63293]: DEBUG nova.virt.hardware [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image limits 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1190.008484] env[63293]: DEBUG nova.virt.hardware [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Flavor pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1190.008634] env[63293]: DEBUG nova.virt.hardware [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Image pref 0:0:0 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1190.008787] env[63293]: DEBUG nova.virt.hardware [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63293) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1190.008999] env[63293]: DEBUG nova.virt.hardware [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1190.009181] env[63293]: DEBUG nova.virt.hardware [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1190.009357] env[63293]: DEBUG nova.virt.hardware [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Got 1 possible topologies {{(pid=63293) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1190.009526] env[63293]: DEBUG nova.virt.hardware [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1190.009701] env[63293]: DEBUG nova.virt.hardware [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63293) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1190.010568] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0129402a-0e0f-40a9-9e21-df1c5e70a743 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.018236] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd734c4-103a-492c-9fa6-80c45294aefe {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.031059] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:6e:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '10af5b0e-12c0-4492-a508-bd47f02b1173', 'vif_model': 'vmxnet3'}] {{(pid=63293) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1190.038471] env[63293]: DEBUG oslo.service.loopingcall [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1190.038696] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Creating VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1190.038895] env[63293]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4fd7b139-55cb-4a99-a474-ba6c5d5157d7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.058266] env[63293]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1190.058266] env[63293]: value = "task-1328339" [ 1190.058266] env[63293]: _type = "Task" [ 1190.058266] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.066567] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328339, 'name': CreateVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.569147] env[63293]: DEBUG oslo_vmware.api [-] Task: {'id': task-1328339, 'name': CreateVM_Task, 'duration_secs': 0.286449} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.569327] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Created VM on the ESX host {{(pid=63293) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1190.569966] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.570152] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.570530] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1190.570782] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c873719c-5bd5-45e1-b875-71ecf345ca5a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.575058] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1190.575058] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5262ecc0-42b2-b1b5-9514-5d7fd03080ff" [ 1190.575058] env[63293]: _type = "Task" [ 1190.575058] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.582187] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]5262ecc0-42b2-b1b5-9514-5d7fd03080ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.721039] env[63293]: DEBUG nova.compute.manager [req-423cff35-863a-42e8-bda0-483cd80839e4 req-e3b6d307-36f4-492b-88fd-4d87bda8d487 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Received event network-changed-10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1190.721248] env[63293]: DEBUG nova.compute.manager [req-423cff35-863a-42e8-bda0-483cd80839e4 req-e3b6d307-36f4-492b-88fd-4d87bda8d487 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Refreshing instance network info cache due to event network-changed-10af5b0e-12c0-4492-a508-bd47f02b1173. {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1190.721467] env[63293]: DEBUG oslo_concurrency.lockutils [req-423cff35-863a-42e8-bda0-483cd80839e4 req-e3b6d307-36f4-492b-88fd-4d87bda8d487 service nova] Acquiring lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.721616] env[63293]: DEBUG oslo_concurrency.lockutils [req-423cff35-863a-42e8-bda0-483cd80839e4 req-e3b6d307-36f4-492b-88fd-4d87bda8d487 service nova] Acquired lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.721776] env[63293]: DEBUG nova.network.neutron [req-423cff35-863a-42e8-bda0-483cd80839e4 req-e3b6d307-36f4-492b-88fd-4d87bda8d487 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Refreshing network info cache for port 10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1191.085061] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.085434] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Processing image 5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1191.085667] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.085822] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.086011] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1191.086268] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-808af68f-618e-4e7b-aca4-ed380d4668b5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.103829] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1191.104038] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63293) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1191.104735] env[63293]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aaa8d3f0-5a17-4c22-960c-1bdb90b382e7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.109769] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1191.109769] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a790c7-0924-8919-a460-1d987297b8de" [ 1191.109769] env[63293]: _type = "Task" [ 1191.109769] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.117270] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52a790c7-0924-8919-a460-1d987297b8de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.415655] env[63293]: DEBUG nova.network.neutron [req-423cff35-863a-42e8-bda0-483cd80839e4 req-e3b6d307-36f4-492b-88fd-4d87bda8d487 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updated VIF entry in instance network info cache for port 10af5b0e-12c0-4492-a508-bd47f02b1173. {{(pid=63293) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1191.416033] env[63293]: DEBUG nova.network.neutron [req-423cff35-863a-42e8-bda0-483cd80839e4 req-e3b6d307-36f4-492b-88fd-4d87bda8d487 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updating instance_info_cache with network_info: [{"id": "10af5b0e-12c0-4492-a508-bd47f02b1173", "address": "fa:16:3e:46:6e:33", "network": {"id": "65e27769-7214-4114-b2ac-098e52910fd6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1083746963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1687c33183e74b6ba70e021b6879cb93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10af5b0e-12", "ovs_interfaceid": "10af5b0e-12c0-4492-a508-bd47f02b1173", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.619794] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Preparing fetch location {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1191.620079] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Fetch image to [datastore2] OSTACK_IMG_207e4181-bc1b-4338-9bef-7e3a9f535b0f/OSTACK_IMG_207e4181-bc1b-4338-9bef-7e3a9f535b0f.vmdk {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1191.620307] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Downloading stream optimized image 5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f to [datastore2] OSTACK_IMG_207e4181-bc1b-4338-9bef-7e3a9f535b0f/OSTACK_IMG_207e4181-bc1b-4338-9bef-7e3a9f535b0f.vmdk on the data store datastore2 as vApp {{(pid=63293) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1191.620453] env[63293]: DEBUG nova.virt.vmwareapi.images [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Downloading image file data 5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f to the ESX as VM named 'OSTACK_IMG_207e4181-bc1b-4338-9bef-7e3a9f535b0f' {{(pid=63293) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1191.686520] env[63293]: DEBUG oslo_vmware.rw_handles [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1191.686520] env[63293]: value = "resgroup-9" [ 1191.686520] env[63293]: _type = "ResourcePool" [ 1191.686520] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1191.686882] env[63293]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-4be674e8-968b-41a8-aa82-0e86d991c147 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.707390] env[63293]: DEBUG oslo_vmware.rw_handles [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lease: (returnval){ [ 1191.707390] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e39c71-7b90-219d-e30d-ddb8777d6ee1" [ 1191.707390] env[63293]: _type = "HttpNfcLease" [ 1191.707390] env[63293]: } obtained for vApp import into resource pool (val){ [ 1191.707390] env[63293]: value = "resgroup-9" [ 1191.707390] env[63293]: _type = "ResourcePool" [ 1191.707390] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1191.707755] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the lease: (returnval){ [ 1191.707755] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e39c71-7b90-219d-e30d-ddb8777d6ee1" [ 1191.707755] env[63293]: _type = "HttpNfcLease" [ 1191.707755] env[63293]: } to be ready. {{(pid=63293) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1191.713727] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1191.713727] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e39c71-7b90-219d-e30d-ddb8777d6ee1" [ 1191.713727] env[63293]: _type = "HttpNfcLease" [ 1191.713727] env[63293]: } is initializing. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1191.867321] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1191.918471] env[63293]: DEBUG oslo_concurrency.lockutils [req-423cff35-863a-42e8-bda0-483cd80839e4 req-e3b6d307-36f4-492b-88fd-4d87bda8d487 service nova] Releasing lock "refresh_cache-7ae46382-8372-4a33-a9b6-e0720f09811e" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.216055] env[63293]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1192.216055] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e39c71-7b90-219d-e30d-ddb8777d6ee1" [ 1192.216055] env[63293]: _type = "HttpNfcLease" [ 1192.216055] env[63293]: } is ready. {{(pid=63293) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1192.216490] env[63293]: DEBUG oslo_vmware.rw_handles [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1192.216490] env[63293]: value = "session[52ade92f-3039-8f34-6f5f-c0f8e492e4bd]52e39c71-7b90-219d-e30d-ddb8777d6ee1" [ 1192.216490] env[63293]: _type = "HttpNfcLease" [ 1192.216490] env[63293]: }. {{(pid=63293) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1192.217081] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c177ba-da4c-4ef4-81e8-c0587ed3cc86 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.224162] env[63293]: DEBUG oslo_vmware.rw_handles [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52822faf-70ce-35c8-d2c9-eefc55487565/disk-0.vmdk from lease info. {{(pid=63293) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1192.224320] env[63293]: DEBUG oslo_vmware.rw_handles [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52822faf-70ce-35c8-d2c9-eefc55487565/disk-0.vmdk. {{(pid=63293) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1192.287072] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b500c823-40c5-4e9b-a2a4-d9faf6f793bb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.871757] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.341490] env[63293]: DEBUG oslo_vmware.rw_handles [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Completed reading data from the image iterator. {{(pid=63293) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1193.341867] env[63293]: DEBUG oslo_vmware.rw_handles [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52822faf-70ce-35c8-d2c9-eefc55487565/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1193.342472] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386905fe-d8e8-4e70-89ab-164ca44067a8 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.349045] env[63293]: DEBUG oslo_vmware.rw_handles [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52822faf-70ce-35c8-d2c9-eefc55487565/disk-0.vmdk is in state: ready. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1193.349243] env[63293]: DEBUG oslo_vmware.rw_handles [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52822faf-70ce-35c8-d2c9-eefc55487565/disk-0.vmdk. {{(pid=63293) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1193.349475] env[63293]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-bfb10cca-9e83-4609-a00f-e24d3f6f879a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.539664] env[63293]: DEBUG oslo_vmware.rw_handles [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52822faf-70ce-35c8-d2c9-eefc55487565/disk-0.vmdk. {{(pid=63293) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1193.539924] env[63293]: INFO nova.virt.vmwareapi.images [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Downloaded image file data 5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f [ 1193.540842] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8045480-a244-4f9f-a527-fd9ca39e8a6a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.557401] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fae366cb-611a-439b-bfe1-44310f6aa22d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.588775] env[63293]: INFO nova.virt.vmwareapi.images [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] The imported VM was unregistered [ 1193.591080] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Caching image {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1193.591324] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Creating directory with path [datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1193.591611] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81f45067-def6-496c-a21b-3a260c1ff94b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.613446] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Created directory with path [datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f {{(pid=63293) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1193.613627] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_207e4181-bc1b-4338-9bef-7e3a9f535b0f/OSTACK_IMG_207e4181-bc1b-4338-9bef-7e3a9f535b0f.vmdk to [datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f.vmdk. {{(pid=63293) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1193.613886] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-fb1a458a-7481-47bc-afcf-0abd5a3adefb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.619935] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1193.619935] env[63293]: value = "task-1328342" [ 1193.619935] env[63293]: _type = "Task" [ 1193.619935] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.627210] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328342, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.871339] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.871523] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.871691] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.871856] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.872073] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1194.130496] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328342, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.375972] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.375972] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.376247] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.376398] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63293) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1194.377496] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9254a5f2-2ec5-4bc1-a45c-1b9dc9448b06 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.386421] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6f71ca-1b3c-4ddf-9e45-b3be78fe71a0 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.402088] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72cf5b3-bf8a-4658-b657-e7bf266b1be5 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.409731] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf0191a-77ae-4663-8f06-f1f4cab6d7bd {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.441629] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181093MB free_disk=132GB free_vcpus=48 pci_devices=None {{(pid=63293) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1194.441856] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.442078] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.632303] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328342, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.132290] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328342, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.469271] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Instance 7ae46382-8372-4a33-a9b6-e0720f09811e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63293) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1195.469584] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1195.469746] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63293) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1195.500771] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb4c1399-9123-4626-87ad-bf336bdec94f {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.509010] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71aae48-75ae-4707-8afa-385670d8c0eb {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.541581] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44300f4-e708-40d9-953b-e34f45d95c6d {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.549655] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-941cbac4-a8c1-46ed-b85c-2a97ddb41856 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.564079] env[63293]: DEBUG nova.compute.provider_tree [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1195.633027] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328342, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.068169] env[63293]: DEBUG nova.scheduler.client.report [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1196.133206] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328342, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.223637} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.133480] env[63293]: INFO nova.virt.vmwareapi.ds_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_207e4181-bc1b-4338-9bef-7e3a9f535b0f/OSTACK_IMG_207e4181-bc1b-4338-9bef-7e3a9f535b0f.vmdk to [datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f.vmdk. [ 1196.133696] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Cleaning up location [datastore2] OSTACK_IMG_207e4181-bc1b-4338-9bef-7e3a9f535b0f {{(pid=63293) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1196.133851] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_207e4181-bc1b-4338-9bef-7e3a9f535b0f {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1196.134120] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c234475c-44b1-483a-bcba-feeca81dbf73 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.140386] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1196.140386] env[63293]: value = "task-1328343" [ 1196.140386] env[63293]: _type = "Task" [ 1196.140386] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.147814] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328343, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.573139] env[63293]: DEBUG nova.compute.resource_tracker [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63293) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1196.573525] env[63293]: DEBUG oslo_concurrency.lockutils [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.131s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.651348] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328343, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034009} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.651549] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1196.651724] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f.vmdk" {{(pid=63293) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1196.651974] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f.vmdk to [datastore2] 7ae46382-8372-4a33-a9b6-e0720f09811e/7ae46382-8372-4a33-a9b6-e0720f09811e.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1196.652240] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b15c3590-a953-4f72-a2d0-b0be31f9194b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.658536] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1196.658536] env[63293]: value = "task-1328344" [ 1196.658536] env[63293]: _type = "Task" [ 1196.658536] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.665723] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328344, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.168876] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328344, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.670812] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328344, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.171269] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328344, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.574381] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1198.574624] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Starting heal instance info cache {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1198.672213] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328344, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.171112] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328344, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.136672} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.171358] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f/5d15b8d0-da2e-4e3d-9b0c-1aeaea0a101f.vmdk to [datastore2] 7ae46382-8372-4a33-a9b6-e0720f09811e/7ae46382-8372-4a33-a9b6-e0720f09811e.vmdk {{(pid=63293) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1199.172149] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9809835e-b2de-4a3c-8b6c-c8a3bf78567b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.193658] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 7ae46382-8372-4a33-a9b6-e0720f09811e/7ae46382-8372-4a33-a9b6-e0720f09811e.vmdk or device None with type streamOptimized {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1199.193908] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aba521b4-56ed-4ac4-8e10-0d015df49561 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.213044] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1199.213044] env[63293]: value = "task-1328345" [ 1199.213044] env[63293]: _type = "Task" [ 1199.213044] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.220569] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328345, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.723189] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328345, 'name': ReconfigVM_Task, 'duration_secs': 0.299012} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.723533] env[63293]: DEBUG nova.virt.vmwareapi.volumeops [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 7ae46382-8372-4a33-a9b6-e0720f09811e/7ae46382-8372-4a33-a9b6-e0720f09811e.vmdk or device None with type streamOptimized {{(pid=63293) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1199.724127] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-179a71f5-4486-4cb8-b267-8d2aebc35b9a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.730580] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1199.730580] env[63293]: value = "task-1328346" [ 1199.730580] env[63293]: _type = "Task" [ 1199.730580] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.737792] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328346, 'name': Rename_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.084417] env[63293]: DEBUG nova.compute.manager [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Didn't find any instances for network info cache update. {{(pid=63293) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1200.240373] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328346, 'name': Rename_Task, 'duration_secs': 0.377013} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.240638] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Powering on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1200.240919] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5940c2c4-dea0-47d4-a8e4-998313b8f44e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.247177] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1200.247177] env[63293]: value = "task-1328347" [ 1200.247177] env[63293]: _type = "Task" [ 1200.247177] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.255500] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.756756] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328347, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.257920] env[63293]: DEBUG oslo_vmware.api [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328347, 'name': PowerOnVM_Task, 'duration_secs': 0.513189} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.258258] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Powered on the VM {{(pid=63293) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1201.354079] env[63293]: DEBUG nova.compute.manager [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Checking state {{(pid=63293) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1201.355012] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05547bb3-01d8-46ff-9b01-ac2661bdf90b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.871648] env[63293]: DEBUG oslo_concurrency.lockutils [None req-563b2b25-267c-4b11-a7da-e41471829b6a tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.727s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.377182] env[63293]: DEBUG oslo_service.periodic_task [None req-4f550709-81b9-4a79-baca-c7c8f1cc2fe9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63293) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.498527] env[63293]: DEBUG oslo_concurrency.lockutils [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "7ae46382-8372-4a33-a9b6-e0720f09811e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.498787] env[63293]: DEBUG oslo_concurrency.lockutils [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.499013] env[63293]: DEBUG oslo_concurrency.lockutils [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.499213] env[63293]: DEBUG oslo_concurrency.lockutils [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.499407] env[63293]: DEBUG oslo_concurrency.lockutils [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.501426] env[63293]: INFO nova.compute.manager [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Terminating instance [ 1202.503165] env[63293]: DEBUG nova.compute.manager [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Start destroying the instance on the hypervisor. {{(pid=63293) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1202.503373] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Destroying instance {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1202.504203] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f65881-4fce-41ac-beda-6d9c4fea2cae {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.511674] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Powering off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1202.511903] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e5a4230e-0006-41d9-967b-6faa0619b0de {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.518077] env[63293]: DEBUG oslo_vmware.api [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1202.518077] env[63293]: value = "task-1328348" [ 1202.518077] env[63293]: _type = "Task" [ 1202.518077] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.526690] env[63293]: DEBUG oslo_vmware.api [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328348, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.027950] env[63293]: DEBUG oslo_vmware.api [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328348, 'name': PowerOffVM_Task, 'duration_secs': 0.188583} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.028361] env[63293]: DEBUG nova.virt.vmwareapi.vm_util [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Powered off the VM {{(pid=63293) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1203.028413] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Unregistering the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1203.028669] env[63293]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-87976723-259c-40fa-8776-76f97e2aaf6a {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.088194] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Unregistered the VM {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1203.088427] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Deleting contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1203.088615] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleting the datastore file [datastore2] 7ae46382-8372-4a33-a9b6-e0720f09811e {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1203.088882] env[63293]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7eea5d00-2042-4a26-b101-27d5f555b128 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.095500] env[63293]: DEBUG oslo_vmware.api [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for the task: (returnval){ [ 1203.095500] env[63293]: value = "task-1328350" [ 1203.095500] env[63293]: _type = "Task" [ 1203.095500] env[63293]: } to complete. {{(pid=63293) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.102829] env[63293]: DEBUG oslo_vmware.api [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328350, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.604909] env[63293]: DEBUG oslo_vmware.api [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Task: {'id': task-1328350, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132171} completed successfully. {{(pid=63293) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.605211] env[63293]: DEBUG nova.virt.vmwareapi.ds_util [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleted the datastore file {{(pid=63293) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1203.605420] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Deleted contents of the VM from datastore datastore2 {{(pid=63293) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1203.605603] env[63293]: DEBUG nova.virt.vmwareapi.vmops [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Instance destroyed {{(pid=63293) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1203.605783] env[63293]: INFO nova.compute.manager [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1203.606040] env[63293]: DEBUG oslo.service.loopingcall [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63293) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1203.606244] env[63293]: DEBUG nova.compute.manager [-] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Deallocating network for instance {{(pid=63293) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1203.606340] env[63293]: DEBUG nova.network.neutron [-] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] deallocate_for_instance() {{(pid=63293) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1204.008500] env[63293]: DEBUG nova.compute.manager [req-83aa0a1c-afc3-42ca-b0e6-b175a740cc94 req-d0348a92-4a44-4f91-b1f4-31d68901e8e0 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Received event network-vif-deleted-10af5b0e-12c0-4492-a508-bd47f02b1173 {{(pid=63293) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1204.008590] env[63293]: INFO nova.compute.manager [req-83aa0a1c-afc3-42ca-b0e6-b175a740cc94 req-d0348a92-4a44-4f91-b1f4-31d68901e8e0 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Neutron deleted interface 10af5b0e-12c0-4492-a508-bd47f02b1173; detaching it from the instance and deleting it from the info cache [ 1204.009079] env[63293]: DEBUG nova.network.neutron [req-83aa0a1c-afc3-42ca-b0e6-b175a740cc94 req-d0348a92-4a44-4f91-b1f4-31d68901e8e0 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1204.487080] env[63293]: DEBUG nova.network.neutron [-] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Updating instance_info_cache with network_info: [] {{(pid=63293) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1204.511230] env[63293]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-743bf1fb-3eef-49a0-ab6f-620ccff0ec7b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.520754] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df039f90-330d-46c6-aff9-088a2b40a295 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.542460] env[63293]: DEBUG nova.compute.manager [req-83aa0a1c-afc3-42ca-b0e6-b175a740cc94 req-d0348a92-4a44-4f91-b1f4-31d68901e8e0 service nova] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Detach interface failed, port_id=10af5b0e-12c0-4492-a508-bd47f02b1173, reason: Instance 7ae46382-8372-4a33-a9b6-e0720f09811e could not be found. {{(pid=63293) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1204.989676] env[63293]: INFO nova.compute.manager [-] [instance: 7ae46382-8372-4a33-a9b6-e0720f09811e] Took 1.38 seconds to deallocate network for instance. [ 1205.495947] env[63293]: DEBUG oslo_concurrency.lockutils [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.496246] env[63293]: DEBUG oslo_concurrency.lockutils [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.496496] env[63293]: DEBUG nova.objects.instance [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lazy-loading 'resources' on Instance uuid 7ae46382-8372-4a33-a9b6-e0720f09811e {{(pid=63293) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1206.030943] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb36feb-7143-43f3-90cb-0220445fd39e {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.038562] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecbfab2e-ed3a-412a-b6f9-3ed0842d7b1b {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.067473] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911efb77-2c0c-47a3-8010-e3f1b4748ac7 {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.074781] env[63293]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32d9a9a-f561-43c9-9252-faa545edd9aa {{(pid=63293) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.088974] env[63293]: DEBUG nova.compute.provider_tree [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed in ProviderTree for provider: 619725c3-6194-4724-94b1-1dd499be55d0 {{(pid=63293) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1206.592148] env[63293]: DEBUG nova.scheduler.client.report [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Inventory has not changed for provider 619725c3-6194-4724-94b1-1dd499be55d0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 132, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63293) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1207.097108] env[63293]: DEBUG oslo_concurrency.lockutils [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.601s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.117735] env[63293]: INFO nova.scheduler.client.report [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Deleted allocations for instance 7ae46382-8372-4a33-a9b6-e0720f09811e [ 1207.625542] env[63293]: DEBUG oslo_concurrency.lockutils [None req-733a0ad2-bb8e-4ba4-9a0d-616cd46c45c5 tempest-ServerActionsTestOtherB-1806802862 tempest-ServerActionsTestOtherB-1806802862-project-member] Lock "7ae46382-8372-4a33-a9b6-e0720f09811e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.127s {{(pid=63293) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}