Feb 23 20:04:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:04:08.888 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:04:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:04:08.889 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:04:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:04:08.889 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:05:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:05:08.889 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:05:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:05:08.890 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:05:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:05:08.890 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:06:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:06:08.890 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:06:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:06:08.891 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:06:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:06:08.891 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:07:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:07:08.892 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:07:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:07:08.894 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:07:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:07:08.894 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:08:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:08:08.894 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:08:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:08:08.895 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:08:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:08:08.895 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:09:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:09:08.896 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:09:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:09:08.896 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:09:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:09:08.896 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:10:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:10:08.897 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:10:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:10:08.898 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:10:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:10:08.898 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:10:58 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:10:58.192 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 23 20:10:58 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:10:58.193 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 23 20:10:59 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:10:59.322 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:10:59 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:10:59.322 28954 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 23 20:10:59 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:10:59.323 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 23 20:10:59 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:10:59.324 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:10:59 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:10:59.325 28850 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 23 20:10:59 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:10:59.325 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 23 20:11:01 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:01.351 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:11:01 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:01.351 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:11:03 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:03.352 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 20:11:03 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:03.352 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 23 20:11:03 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:03.354 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 20:11:03 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:03.354 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 23 20:11:07 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:07.368 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:11:07 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:07.368 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:11:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:08.899 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:11:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:08.899 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:11:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:08.899 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:11:10 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:10.420 28954 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (No route to host) Feb 23 20:11:10 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:10.420 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 23 20:11:10 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:10.423 28850 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (No route to host) Feb 23 20:11:10 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:10.423 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 23 20:11:42 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:42.544 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 23 20:11:42 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:42.544 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 23 20:11:42 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:42.573 28850 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': '2e:d4:6c', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': '6e:ee:58:9c:24:ba'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Feb 23 20:11:42 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:42.575 28850 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 0 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Feb 23 20:11:42 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:11:42.576 28850 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=7ed7ba5d-4916-47e5-8f3d-b9fae6ee411c, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Feb 23 20:12:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:12:08.901 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:12:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:12:08.902 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:12:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:12:08.902 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:13:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:13:08.902 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:13:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:13:08.903 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:13:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:13:08.903 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:14:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:14:08.904 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:14:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:14:08.904 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:14:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:14:08.904 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:15:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:15:08.904 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:15:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:15:08.905 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:15:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:15:08.905 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:16:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:16:08.906 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:16:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:16:08.907 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:16:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:16:08.907 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:17:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:17:08.907 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:17:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:17:08.908 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:17:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:17:08.908 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:18:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:18:08.909 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:18:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:18:08.910 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:18:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:18:08.910 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:18:59 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:18:59.393 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 23 20:18:59 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:18:59.399 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 23 20:19:00 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:00.402 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:19:00 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:00.404 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:19:01 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:01.404 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 20:19:01 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:01.404 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 23 20:19:01 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:01.406 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 20:19:01 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:01.406 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 23 20:19:03 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:03.414 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:19:03 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:03.436 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:19:05 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:05.416 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 20:19:05 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:05.416 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 23 20:19:05 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:05.438 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 23 20:19:05 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:05.438 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 23 20:19:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:08.911 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:19:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:08.912 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:19:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:08.913 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:19:09 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:09.434 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:19:09 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:09.435 28954 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 23 20:19:09 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:09.436 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 23 20:19:09 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:09.445 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 23 20:19:09 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:09.446 28850 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 23 20:19:09 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:19:09.447 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 23 20:20:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:20:08.913 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:20:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:20:08.913 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:20:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:20:08.914 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:20:37 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:20:37.590 28954 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 23 20:20:37 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:20:37.594 28850 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 23 20:20:37 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:20:37.609 28850 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': '2e:d4:6c', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': '6e:ee:58:9c:24:ba'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Feb 23 20:20:37 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:20:37.610 28850 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 9 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Feb 23 20:20:46 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:20:46.612 28850 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=7ed7ba5d-4916-47e5-8f3d-b9fae6ee411c, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Feb 23 20:20:46 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:20:46.613 28850 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129 Feb 23 20:21:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:21:08.915 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:21:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:21:08.915 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:21:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:21:08.915 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:22:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:22:08.916 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:22:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:22:08.917 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:22:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:22:08.917 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:23:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:23:08.917 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:23:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:23:08.918 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:23:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:23:08.918 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:24:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:08.918 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 23 20:24:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:08.918 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 23 20:24:08 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:08.918 28850 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 23 20:24:20 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:20.827 28954 INFO eventlet.wsgi.server [-] (28954) wsgi exited, is_accepting=True Feb 23 20:24:20 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:20.829 28954 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Feb 23 20:24:20 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:20.835 28954 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Feb 23 20:24:20 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:20.836 28954 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Feb 23 20:24:20 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:20.842 28850 DEBUG oslo_privsep.comm [-] EOF on privsep read channel _reader_main /usr/lib/python3.9/site-packages/oslo_privsep/comm.py:170 Feb 23 20:24:21 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:21.467 28850 INFO oslo_service.service [-] Caught SIGTERM, stopping children Feb 23 20:24:21 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:21.468 28850 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Feb 23 20:24:21 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:21.468 28850 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Feb 23 20:24:21 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:21.469 28850 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Feb 23 20:24:21 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:21.469 28850 DEBUG oslo_service.service [-] Stop services. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:695 Feb 23 20:24:21 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:21.469 28850 DEBUG oslo_service.service [-] Killing children. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:700 Feb 23 20:24:21 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:21.470 28850 INFO oslo_service.service [-] Waiting on 1 children to exit Feb 23 20:24:21 edpm-compute-0 ovn_metadata_agent[28834]: 2026-02-23 20:24:21.470 28850 INFO oslo_service.service [-] Child 28954 exited with status 0 Feb 23 20:24:22 edpm-compute-0 podman[174705]: Error: no container with ID 6a0ee6a0a50d70a628713ac8d8973510b78363bb1342cd671d7270a4e8cd2985 found in database: no such container Feb 23 20:24:22 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Feb 23 20:24:22 edpm-compute-0 podman[174721]: Error: no container with name or ID "ovn_metadata_agent" found: no such container Feb 23 20:24:22 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Feb 23 20:24:22 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Failed with result 'exit-code'. Feb 23 20:24:23 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Scheduled restart job, restart counter is at 1. Feb 23 20:24:23 edpm-compute-0 systemd[1]: Stopped ovn_metadata_agent container. Feb 23 20:24:23 edpm-compute-0 systemd[1]: Starting ovn_metadata_agent container... Feb 23 20:24:23 edpm-compute-0 edpm-start-podman-container[174748]: ovn_metadata_agent Feb 23 20:24:23 edpm-compute-0 edpm-start-podman-container[174747]: Creating additional drop-in dependency for "ovn_metadata_agent" (4213ce4046e52cd3ccf71edca348e36927b796e955d1c2cb1ce8945e253b6805) Feb 23 20:24:24 edpm-compute-0 systemd[1]: Started ovn_metadata_agent container.