Mar 16 08:59:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 08:59:35.254 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 08:59:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 08:59:35.255 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 08:59:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 08:59:35.255 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:00:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:00:35.256 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:00:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:00:35.257 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:00:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:00:35.258 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:01:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:01:35.257 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:01:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:01:35.258 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:01:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:01:35.258 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:02:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:02:35.259 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:02:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:02:35.259 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:02:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:02:35.259 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:03:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:03:35.260 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:03:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:03:35.260 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:03:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:03:35.261 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:04:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:04:35.260 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:04:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:04:35.260 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:04:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:04:35.261 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:05:27 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:27.838 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Mar 16 09:05:27 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:27.839 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Mar 16 09:05:28 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:28.850 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:05:28 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:28.852 28973 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Mar 16 09:05:28 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:28.852 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Mar 16 09:05:28 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:28.861 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:05:28 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:28.863 28874 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Mar 16 09:05:28 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:28.863 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Mar 16 09:05:30 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:30.885 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:05:30 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:30.888 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:05:32 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:32.886 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 09:05:32 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:32.886 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Mar 16 09:05:32 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:32.891 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 09:05:32 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:32.891 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Mar 16 09:05:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:35.261 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:05:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:35.261 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:05:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:35.262 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:05:36 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:36.895 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:05:36 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:36.898 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:05:40 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:40.896 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 09:05:40 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:40.897 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Mar 16 09:05:40 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:40.901 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 09:05:40 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:05:40.901 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Mar 16 09:06:24 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:06:24.141 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Mar 16 09:06:24 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:06:24.161 28874 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': 'd6:c3:11', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'c6:15:49:ab:9d:d6'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Mar 16 09:06:24 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:06:24.161 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Mar 16 09:06:24 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:06:24.162 28874 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 0 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Mar 16 09:06:24 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:06:24.180 28874 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=a8a47e55-01ae-4a29-9c93-497fa5e96615, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Mar 16 09:06:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:06:35.263 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:06:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:06:35.264 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:06:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:06:35.264 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:07:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:07:35.265 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:07:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:07:35.265 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:07:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:07:35.266 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:08:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:08:35.265 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:08:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:08:35.266 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:08:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:08:35.266 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:09:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:09:35.266 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:09:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:09:35.267 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:09:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:09:35.268 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:10:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:10:35.267 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:10:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:10:35.268 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:10:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:10:35.269 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:11:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:11:35.270 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:11:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:11:35.271 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:11:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:11:35.271 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:12:26 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:26.316 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Mar 16 09:12:26 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:26.318 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Mar 16 09:12:27 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:27.356 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:12:27 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:27.358 28874 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Mar 16 09:12:27 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:27.358 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Mar 16 09:12:27 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:27.360 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:12:27 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:27.361 28973 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Mar 16 09:12:27 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:27.362 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Mar 16 09:12:29 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:29.366 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:12:29 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:29.367 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:12:31 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:31.368 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 09:12:31 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:31.369 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Mar 16 09:12:31 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:31.369 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Mar 16 09:12:31 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:31.369 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Mar 16 09:12:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:35.273 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:12:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:35.274 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:12:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:35.274 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:12:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:35.390 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:12:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:35.392 28973 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Mar 16 09:12:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:35.393 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Mar 16 09:12:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:35.394 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Mar 16 09:12:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:35.395 28874 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Mar 16 09:12:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:12:35.396 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Mar 16 09:13:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:13:35.273 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:13:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:13:35.274 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:13:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:13:35.274 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:14:11 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:14:11.554 28874 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Mar 16 09:14:11 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:14:11.566 28973 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Mar 16 09:14:11 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:14:11.579 28874 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': 'd6:c3:11', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'c6:15:49:ab:9d:d6'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Mar 16 09:14:11 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:14:11.582 28874 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 6 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Mar 16 09:14:17 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:14:17.586 28874 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=a8a47e55-01ae-4a29-9c93-497fa5e96615, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Mar 16 09:14:17 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:14:17.587 28874 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129 Mar 16 09:14:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:14:35.275 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:14:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:14:35.276 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:14:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:14:35.276 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:15:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:15:35.276 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:15:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:15:35.277 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:15:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:15:35.277 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:16:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:16:35.277 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:16:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:16:35.279 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:16:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:16:35.279 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:17:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:35.278 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Mar 16 09:17:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:35.278 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Mar 16 09:17:35 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:35.279 28874 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Mar 16 09:17:43 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:43.715 28973 INFO eventlet.wsgi.server [-] (28973) wsgi exited, is_accepting=True Mar 16 09:17:43 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:43.716 28973 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Mar 16 09:17:43 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:43.716 28973 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Mar 16 09:17:43 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:43.716 28973 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Mar 16 09:17:43 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:43.718 28874 DEBUG oslo_privsep.comm [-] EOF on privsep read channel _reader_main /usr/lib/python3.9/site-packages/oslo_privsep/comm.py:170 Mar 16 09:17:44 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:44.046 28874 INFO oslo_service.service [-] Caught SIGTERM, stopping children Mar 16 09:17:44 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:44.047 28874 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Mar 16 09:17:44 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:44.048 28874 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Mar 16 09:17:44 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:44.048 28874 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Mar 16 09:17:44 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:44.048 28874 DEBUG oslo_service.service [-] Stop services. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:695 Mar 16 09:17:44 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:44.049 28874 DEBUG oslo_service.service [-] Killing children. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:700 Mar 16 09:17:44 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:44.049 28874 INFO oslo_service.service [-] Waiting on 1 children to exit Mar 16 09:17:44 edpm-compute-0 ovn_metadata_agent[28869]: 2026-03-16 09:17:44.050 28874 INFO oslo_service.service [-] Child 28973 exited with status 0 Mar 16 09:17:46 edpm-compute-0 podman[174728]: Error: no container with ID da1cac4ec23562edc1e11321e9318b0c685c35f8f33f686c98ee9b2d7c9c8259 found in database: no such container Mar 16 09:17:46 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Mar 16 09:17:46 edpm-compute-0 podman[174744]: Error: no container with name or ID "ovn_metadata_agent" found: no such container Mar 16 09:17:46 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Mar 16 09:17:46 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Failed with result 'exit-code'. Mar 16 09:17:46 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Scheduled restart job, restart counter is at 1. Mar 16 09:17:46 edpm-compute-0 systemd[1]: Stopped ovn_metadata_agent container. Mar 16 09:17:46 edpm-compute-0 systemd[1]: Starting ovn_metadata_agent container... Mar 16 09:17:46 edpm-compute-0 edpm-start-podman-container[174771]: ovn_metadata_agent Mar 16 09:17:47 edpm-compute-0 edpm-start-podman-container[174769]: Creating additional drop-in dependency for "ovn_metadata_agent" (d01b41ab4c96de5633889165709430c637d9265dbf4103ea84fa87b57f13c1f2) Mar 16 09:17:47 edpm-compute-0 systemd[1]: Started ovn_metadata_agent container.