Mar 16 16:36:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:36:48.789 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:36:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:36:48.791 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:36:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:36:48.791 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:37:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:37:48.791 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:37:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:37:48.793 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:37:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:37:48.793 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:38:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:38:48.791 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:38:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:38:48.792 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:38:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:38:48.792 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:39:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:39:48.792 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:39:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:39:48.793 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:39:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:39:48.793 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:40:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:40:48.793 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:40:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:40:48.793 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:40:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:40:48.793 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:41:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:41:48.864 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:41:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:41:48.865 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:41:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:41:48.865 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:42:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:42:48.865 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:42:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:42:48.866 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:42:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:42:48.866 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:43:01 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:01.138 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Mar 16 16:43:01 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:01.139 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Mar 16 16:43:02 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:02.144 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:43:02 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:02.145 28952 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Mar 16 16:43:02 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:02.145 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Mar 16 16:43:02 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:02.149 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:43:02 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:02.149 28855 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Mar 16 16:43:02 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:02.149 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Mar 16 16:43:04 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:04.157 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:43:04 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:04.157 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:43:06 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:06.164 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 16:43:06 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:06.164 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 16:43:06 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:06.165 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Mar 16 16:43:06 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:06.164 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Mar 16 16:43:10 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:10.182 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:43:10 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:10.184 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:43:14 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:14.186 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 16:43:14 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:14.187 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Mar 16 16:43:14 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:14.187 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 16:43:14 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:14.187 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Mar 16 16:43:30 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:30.221 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Mar 16 16:43:30 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:30.243 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Mar 16 16:43:30 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:30.260 28855 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': 'b2:09:0a', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': '86:ef:43:0e:dc:bd'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Mar 16 16:43:30 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:30.261 28855 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 0 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Mar 16 16:43:30 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:30.262 28855 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=2578fdfb-9e9c-46bc-ac11-5f2cab3e085a, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Mar 16 16:43:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:48.867 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:43:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:48.868 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:43:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:43:48.868 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:44:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:44:48.870 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:44:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:44:48.870 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:44:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:44:48.870 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:45:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:45:48.871 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:45:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:45:48.872 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:45:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:45:48.872 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:46:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:46:48.872 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:46:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:46:48.872 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:46:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:46:48.872 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:47:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:47:48.873 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:47:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:47:48.873 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:47:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:47:48.873 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:48:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:48:48.874 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:48:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:48:48.876 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:48:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:48:48.876 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:49:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:48.159 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Mar 16 16:49:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:48.160 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Mar 16 16:49:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:48.876 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:49:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:48.876 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:49:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:48.876 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:49:49 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:49.171 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:49:49 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:49.171 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:49:50 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:50.173 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 16:49:50 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:50.173 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Mar 16 16:49:50 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:50.173 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 16:49:50 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:50.174 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Mar 16 16:49:52 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:52.183 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:49:52 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:52.185 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:49:54 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:54.185 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 16:49:54 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:54.185 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Mar 16 16:49:54 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:54.188 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 16:49:54 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:54.188 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Mar 16 16:49:58 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:58.199 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:49:58 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:58.200 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 16:49:58 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:58.202 28855 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Mar 16 16:49:58 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:58.202 28952 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Mar 16 16:49:58 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:58.203 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Mar 16 16:49:58 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:49:58.202 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Mar 16 16:50:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:50:48.877 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:50:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:50:48.877 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:50:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:50:48.878 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:51:18 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:51:18.308 28952 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Mar 16 16:51:18 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:51:18.313 28855 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Mar 16 16:51:18 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:51:18.323 28855 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': 'b2:09:0a', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': '86:ef:43:0e:dc:bd'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Mar 16 16:51:18 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:51:18.324 28855 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 4 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Mar 16 16:51:22 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:51:22.326 28855 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=2578fdfb-9e9c-46bc-ac11-5f2cab3e085a, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Mar 16 16:51:22 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:51:22.328 28855 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129 Mar 16 16:51:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:51:48.878 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:51:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:51:48.878 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:51:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:51:48.878 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:52:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:52:48.878 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:52:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:52:48.879 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:52:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:52:48.879 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:53:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:53:48.879 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:53:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:53:48.880 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:53:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:53:48.881 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:54:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:54:48.880 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 16:54:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:54:48.881 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 16:54:48 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:54:48.881 28855 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.665 28952 INFO eventlet.wsgi.server [-] (28952) wsgi exited, is_accepting=True Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.666 28952 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.666 28952 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.667 28952 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.667 28855 DEBUG oslo_privsep.comm [-] EOF on privsep read channel _reader_main /usr/lib/python3.9/site-packages/oslo_privsep/comm.py:170 Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.801 28855 INFO oslo_service.service [-] Caught SIGTERM, stopping children Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.802 28855 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.802 28855 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.803 28855 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.804 28855 DEBUG oslo_service.service [-] Stop services. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:695 Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.805 28855 DEBUG oslo_service.service [-] Killing children. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:700 Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.806 28855 INFO oslo_service.service [-] Waiting on 1 children to exit Mar 16 16:55:23 edpm-compute-0 ovn_metadata_agent[28850]: 2026-03-16 16:55:23.806 28855 INFO oslo_service.service [-] Child 28952 exited with status 0 Mar 16 16:55:27 edpm-compute-0 podman[174572]: Error: no container with ID 0154b7e1b15355bcd308dce01e1c55acd535292c95f6e793458ea1af13ab9b8a found in database: no such container Mar 16 16:55:27 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Mar 16 16:55:27 edpm-compute-0 podman[174587]: Error: no container with name or ID "ovn_metadata_agent" found: no such container Mar 16 16:55:27 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Mar 16 16:55:27 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Failed with result 'exit-code'. Mar 16 16:55:27 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Scheduled restart job, restart counter is at 1. Mar 16 16:55:27 edpm-compute-0 systemd[1]: Stopped ovn_metadata_agent container. Mar 16 16:55:27 edpm-compute-0 systemd[1]: Starting ovn_metadata_agent container... Mar 16 16:55:27 edpm-compute-0 edpm-start-podman-container[174618]: ovn_metadata_agent Mar 16 16:55:27 edpm-compute-0 edpm-start-podman-container[174617]: Creating additional drop-in dependency for "ovn_metadata_agent" (c47b7085c87b741470a479686b09d6e71ede01aacb2ee0022c142e2db082f1bd) Mar 16 16:55:27 edpm-compute-0 systemd[1]: Started ovn_metadata_agent container.