Feb 23 10:12:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:12:00.163 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:12:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:12:00.163 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:12:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:12:00.163 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:13:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:13:00.163 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:13:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:13:00.164 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:13:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:13:00.164 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:14:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:14:00.164 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:14:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:14:00.165 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:14:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:14:00.165 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:15:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:15:00.164 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:15:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:15:00.165 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:15:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:15:00.165 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:16:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:16:00.166 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:16:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:16:00.167 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:16:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:16:00.167 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:17:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:17:00.168 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:17:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:17:00.169 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:17:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:17:00.169 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:18:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:00.169 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:18:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:00.170 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:18:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:00.170 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:18:50 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:50.016 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 23 10:18:50 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:50.019 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 23 10:18:51 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:51.071 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:18:51 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:51.073 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:18:51 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:51.074 28960 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 23 10:18:51 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:51.075 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 23 10:18:51 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:51.075 28863 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 23 10:18:51 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:51.075 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 23 10:18:53 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:53.107 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:18:53 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:53.110 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:18:55 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:55.110 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 10:18:55 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:55.111 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 23 10:18:55 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:55.112 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 10:18:55 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:55.112 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 23 10:18:59 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:59.122 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:18:59 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:18:59.122 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:19:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:00.171 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:19:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:00.171 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:19:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:00.172 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:19:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:03.124 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 10:19:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:03.125 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 23 10:19:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:03.126 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 10:19:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:03.126 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 23 10:19:19 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:19.214 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 23 10:19:19 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:19.219 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 23 10:19:19 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:19.241 28863 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': 'e2:1d:c0', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'ee:45:5d:45:ff:b0'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Feb 23 10:19:19 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:19.243 28863 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 0 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Feb 23 10:19:19 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:19:19.245 28863 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=172380d5-dc0e-4c08-b7da-a1ecc5c63b44, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Feb 23 10:20:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:20:00.173 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:20:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:20:00.173 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:20:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:20:00.174 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:21:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:21:00.175 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:21:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:21:00.175 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:21:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:21:00.176 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:22:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:22:00.177 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:22:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:22:00.178 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:22:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:22:00.178 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:23:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:23:00.178 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:23:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:23:00.179 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:23:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:23:00.179 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:24:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:24:00.180 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:24:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:24:00.180 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:24:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:24:00.181 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:25:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:25:00.181 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:25:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:25:00.182 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:25:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:25:00.182 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:26:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:00.183 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:26:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:00.184 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:26:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:00.184 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:26:52 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:52.859 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 23 10:26:52 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:52.860 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 23 10:26:53 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:53.949 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:26:53 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:53.949 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:26:54 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:54.951 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 10:26:54 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:54.951 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 10:26:54 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:54.951 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 23 10:26:54 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:54.951 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 23 10:26:57 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:57.049 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:26:57 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:57.050 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:26:59 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:59.051 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 10:26:59 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:59.052 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 23 10:26:59 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:59.053 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 10:26:59 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:26:59.053 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 23 10:27:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:27:00.184 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:27:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:27:00.184 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:27:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:27:00.185 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:27:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:27:03.079 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:27:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:27:03.081 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 10:27:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:27:03.084 28863 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 23 10:27:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:27:03.084 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 23 10:27:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:27:03.084 28960 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 23 10:27:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:27:03.085 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 23 10:28:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:28:00.186 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:28:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:28:00.186 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:28:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:28:00.186 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:29:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:29:00.187 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:29:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:29:00.187 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:29:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:29:00.187 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:29:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:29:03.495 28863 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 23 10:29:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:29:03.503 28863 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': 'e2:1d:c0', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'ee:45:5d:45:ff:b0'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Feb 23 10:29:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:29:03.504 28863 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 9 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Feb 23 10:29:03 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:29:03.540 28960 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 23 10:29:12 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:29:12.507 28863 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=172380d5-dc0e-4c08-b7da-a1ecc5c63b44, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Feb 23 10:29:12 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:29:12.507 28863 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129 Feb 23 10:30:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:30:00.189 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:30:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:30:00.189 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:30:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:30:00.189 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:31:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:31:00.190 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:31:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:31:00.190 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:31:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:31:00.190 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:32:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:00.191 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 10:32:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:00.191 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 10:32:00 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:00.192 28863 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.456 28960 INFO eventlet.wsgi.server [-] (28960) wsgi exited, is_accepting=True Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.456 28960 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.457 28960 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.457 28960 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.462 28863 DEBUG oslo_privsep.comm [-] EOF on privsep read channel _reader_main /usr/lib/python3.9/site-packages/oslo_privsep/comm.py:170 Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.732 28863 INFO oslo_service.service [-] Caught SIGTERM, stopping children Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.733 28863 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.734 28863 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.734 28863 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.735 28863 DEBUG oslo_service.service [-] Stop services. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:695 Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.735 28863 DEBUG oslo_service.service [-] Killing children. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:700 Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.736 28863 INFO oslo_service.service [-] Waiting on 1 children to exit Feb 23 10:32:38 edpm-compute-0 ovn_metadata_agent[28858]: 2026-02-23 10:32:38.736 28863 INFO oslo_service.service [-] Child 28960 exited with status 0 Feb 23 10:32:41 edpm-compute-0 podman[174985]: Error: no container with ID e7a1476494118fcdb21cb7b7f24100845aeb85a2aefafb28c49aacf10ad59d01 found in database: no such container Feb 23 10:32:41 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Feb 23 10:32:41 edpm-compute-0 podman[175000]: Error: no container with name or ID "ovn_metadata_agent" found: no such container Feb 23 10:32:41 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Feb 23 10:32:41 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Failed with result 'exit-code'. Feb 23 10:32:41 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Scheduled restart job, restart counter is at 1. Feb 23 10:32:41 edpm-compute-0 systemd[1]: Stopped ovn_metadata_agent container. Feb 23 10:32:41 edpm-compute-0 systemd[1]: Starting ovn_metadata_agent container... Feb 23 10:32:41 edpm-compute-0 edpm-start-podman-container[175030]: ovn_metadata_agent Feb 23 10:32:41 edpm-compute-0 edpm-start-podman-container[175027]: Creating additional drop-in dependency for "ovn_metadata_agent" (3238e2a3d54d2515021e4353e8ee59c4fe943ae9ba8bd15d20ce48825b83131a) Feb 23 10:32:42 edpm-compute-0 systemd[1]: Started ovn_metadata_agent container.