Feb 26 14:36:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:36:05.751 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:36:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:36:05.752 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:36:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:36:05.752 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:37:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:37:05.752 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:37:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:37:05.753 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:37:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:37:05.753 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:38:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:38:05.754 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:38:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:38:05.755 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:38:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:38:05.755 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:39:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:39:05.755 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:39:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:39:05.755 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:39:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:39:05.755 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:40:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:40:05.756 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:40:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:40:05.757 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:40:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:40:05.757 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:41:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:41:05.758 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:41:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:41:05.758 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:41:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:41:05.758 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:42:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:05.758 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:42:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:05.759 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:42:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:05.759 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:42:43 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:43.982 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 26 14:42:43 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:43.983 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 26 14:42:45 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:45.014 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:42:45 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:45.015 28942 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 26 14:42:45 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:45.015 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 26 14:42:45 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:45.016 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:42:45 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:45.016 28826 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 26 14:42:45 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:45.017 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 26 14:42:47 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:47.026 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:42:47 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:47.027 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:42:49 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:49.028 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 26 14:42:49 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:49.029 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 26 14:42:49 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:49.029 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 26 14:42:49 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:49.030 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 26 14:42:53 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:53.047 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:42:53 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:53.051 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:42:57 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:57.052 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 26 14:42:57 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:57.052 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 26 14:42:57 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:57.053 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 26 14:42:57 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:42:57.054 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 26 14:43:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:43:05.759 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:43:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:43:05.760 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:43:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:43:05.760 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:43:38 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:43:38.759 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 26 14:43:38 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:43:38.765 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 26 14:43:38 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:43:38.777 28826 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': '2a:ae:70', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': '56:0e:7f:ec:4e:99'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Feb 26 14:43:38 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:43:38.779 28826 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 0 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Feb 26 14:43:38 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:43:38.780 28826 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=35e59764-7444-420e-a59d-7e0e3caf84bc, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Feb 26 14:44:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:44:05.762 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:44:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:44:05.762 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:44:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:44:05.762 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:45:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:45:05.764 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:45:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:45:05.764 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:45:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:45:05.765 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:46:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:46:05.766 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:46:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:46:05.766 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:46:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:46:05.766 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:47:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:47:05.766 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:47:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:47:05.766 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:47:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:47:05.766 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:48:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:48:05.767 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:48:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:48:05.768 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:48:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:48:05.768 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:49:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:49:05.769 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:49:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:49:05.769 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:49:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:49:05.770 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:50:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:50:05.769 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:50:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:50:05.770 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:50:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:50:05.770 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:51:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:05.771 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:51:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:05.771 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:51:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:05.771 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:51:07 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:07.172 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 26 14:51:07 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:07.172 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 26 14:51:08 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:08.212 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:51:08 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:08.213 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:51:09 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:09.214 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 26 14:51:09 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:09.214 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 26 14:51:09 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:09.215 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 26 14:51:09 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:09.216 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 26 14:51:11 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:11.233 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:51:11 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:11.234 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:51:13 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:13.235 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 26 14:51:13 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:13.235 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 26 14:51:13 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:13.236 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 26 14:51:13 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:13.236 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 26 14:51:17 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:17.271 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:51:17 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:17.271 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 26 14:51:17 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:17.273 28942 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 26 14:51:17 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:17.273 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 26 14:51:17 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:17.273 28826 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 26 14:51:17 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:51:17.274 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 26 14:52:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:52:05.772 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:52:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:52:05.772 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:52:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:52:05.772 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:52:45 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:52:45.379 28826 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 26 14:52:45 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:52:45.398 28942 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 26 14:52:45 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:52:45.401 28826 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': '2a:ae:70', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': '56:0e:7f:ec:4e:99'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Feb 26 14:52:45 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:52:45.403 28826 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 2 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Feb 26 14:52:47 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:52:47.406 28826 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=35e59764-7444-420e-a59d-7e0e3caf84bc, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Feb 26 14:52:47 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:52:47.407 28826 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129 Feb 26 14:53:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:53:05.773 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:53:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:53:05.773 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:53:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:53:05.773 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:54:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:54:05.774 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:54:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:54:05.775 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:54:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:54:05.775 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:55:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:55:05.776 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:55:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:55:05.777 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:55:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:55:05.777 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:56:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:05.777 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 14:56:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:05.777 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 14:56:05 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:05.777 28826 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.620 28942 INFO eventlet.wsgi.server [-] (28942) wsgi exited, is_accepting=True Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.620 28826 DEBUG oslo_privsep.comm [-] EOF on privsep read channel _reader_main /usr/lib/python3.9/site-packages/oslo_privsep/comm.py:170 Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.621 28942 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.621 28942 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.621 28942 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.912 28826 INFO oslo_service.service [-] Caught SIGTERM, stopping children Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.913 28826 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.913 28826 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.914 28826 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.914 28826 DEBUG oslo_service.service [-] Stop services. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:695 Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.915 28826 DEBUG oslo_service.service [-] Killing children. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:700 Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.915 28826 INFO oslo_service.service [-] Waiting on 1 children to exit Feb 26 14:56:36 edpm-compute-0 ovn_metadata_agent[28821]: 2026-02-26 14:56:36.915 28826 INFO oslo_service.service [-] Child 28942 exited with status 0 Feb 26 14:56:40 edpm-compute-0 podman[174812]: Error: no container with ID 21fa81250098d52e0900e009dc1e011d440db7442b28aec29e50045c2f5dc733 found in database: no such container Feb 26 14:56:40 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Feb 26 14:56:40 edpm-compute-0 podman[174827]: Error: no container with name or ID "ovn_metadata_agent" found: no such container Feb 26 14:56:40 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Feb 26 14:56:40 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Failed with result 'exit-code'. Feb 26 14:56:40 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Scheduled restart job, restart counter is at 1. Feb 26 14:56:40 edpm-compute-0 systemd[1]: Stopped ovn_metadata_agent container. Feb 26 14:56:40 edpm-compute-0 systemd[1]: Starting ovn_metadata_agent container... Feb 26 14:56:40 edpm-compute-0 edpm-start-podman-container[174859]: ovn_metadata_agent Feb 26 14:56:40 edpm-compute-0 edpm-start-podman-container[174858]: Creating additional drop-in dependency for "ovn_metadata_agent" (e15a95e3091eeca86e4543616759ab325abf84101e31a986b958d375aa6a757b) Feb 26 14:56:41 edpm-compute-0 systemd[1]: Started ovn_metadata_agent container.