Feb 25 23:43:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:43:27.715 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:43:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:43:27.716 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:43:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:43:27.716 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:44:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:44:27.715 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:44:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:44:27.716 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:44:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:44:27.716 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:45:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:45:27.717 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:45:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:45:27.721 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.004s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:45:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:45:27.721 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:46:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:46:27.719 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:46:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:46:27.719 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:46:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:46:27.719 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:47:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:47:27.720 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:47:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:47:27.721 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:47:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:47:27.721 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:48:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:48:27.721 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:48:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:48:27.721 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:48:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:48:27.722 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:49:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:49:27.722 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:49:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:49:27.723 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:49:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:49:27.724 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:50:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:27.724 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:50:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:27.724 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:50:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:27.724 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:50:28 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:28.997 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 25 23:50:28 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:28.998 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 25 23:50:30 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:30.051 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:50:30 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:30.052 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:50:30 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:30.053 28859 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 25 23:50:30 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:30.053 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 25 23:50:30 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:30.054 28953 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 25 23:50:30 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:30.054 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 25 23:50:32 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:32.067 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:50:32 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:32.068 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:50:34 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:34.069 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 25 23:50:34 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:34.069 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 25 23:50:34 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:34.071 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 25 23:50:34 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:34.071 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 25 23:50:38 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:38.107 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:50:38 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:38.109 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:50:42 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:42.109 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 25 23:50:42 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:42.110 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 25 23:50:42 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:42.113 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 25 23:50:42 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:42.113 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 25 23:50:50 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:50.135 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 25 23:50:50 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:50.137 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 25 23:50:50 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:50.148 28859 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': '92:5f:0d', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'e6:fe:01:d2:08:49'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Feb 25 23:50:50 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:50.150 28859 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 0 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Feb 25 23:50:50 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:50:50.152 28859 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=5cf4be0c-ece4-4570-a909-9a606b819258, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Feb 25 23:51:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:51:27.725 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:51:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:51:27.725 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:51:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:51:27.725 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:52:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:52:27.726 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:52:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:52:27.727 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:52:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:52:27.727 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:53:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:53:27.727 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:53:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:53:27.728 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:53:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:53:27.728 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:54:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:54:27.727 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:54:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:54:27.728 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:54:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:54:27.728 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:55:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:55:27.729 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:55:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:55:27.729 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:55:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:55:27.730 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:56:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:56:27.731 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:56:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:56:27.732 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:56:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:56:27.732 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:57:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:57:27.733 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:57:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:57:27.734 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:57:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:57:27.734 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:58:07 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:07.838 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 25 23:58:07 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:07.838 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection closed by peer Feb 25 23:58:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:08.875 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:58:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:08.880 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:58:09 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:09.877 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 25 23:58:09 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:09.878 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 25 23:58:09 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:09.882 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 25 23:58:09 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:09.883 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 2 seconds before reconnect Feb 25 23:58:11 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:11.887 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:58:11 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:11.905 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:58:13 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:13.889 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 25 23:58:13 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:13.889 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 25 23:58:13 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:13.908 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt timed out Feb 25 23:58:13 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:13.909 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: waiting 4 seconds before reconnect Feb 25 23:58:17 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:17.897 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:58:17 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:17.899 28859 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 25 23:58:17 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:17.900 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 25 23:58:17 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:17.918 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connecting... Feb 25 23:58:17 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:17.918 28953 WARNING ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connection attempt failed (Connection refused) Feb 25 23:58:17 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:17.918 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: continuing to reconnect in the background but suppressing further logging Feb 25 23:58:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:27.735 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:58:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:27.737 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.002s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:58:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:58:27.737 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:59:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:59:27.736 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 25 23:59:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:59:27.736 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 25 23:59:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:59:27.737 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 25 23:59:54 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:59:54.031 28859 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 25 23:59:54 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:59:54.094 28859 DEBUG ovsdbapp.backend.ovs_idl.event [-] Matched UPDATE: SbGlobalUpdateEvent(events=('update',), table='SB_Global', conditions=None, old_conditions=None), priority=20 to row=SB_Global(external_ids={}, nb_cfg=1, options={'arp_ns_explicit_output': 'true', 'mac_prefix': '92:5f:0d', 'max_tunid': '16711680', 'northd_internal_version': '24.03.8-20.33.0-76.8', 'svc_monitor_mac': 'e6:fe:01:d2:08:49'}, ipsec=False) old=SB_Global() matches /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/event.py:43 Feb 25 23:59:54 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:59:54.097 28859 DEBUG neutron.agent.ovn.metadata.agent [-] Delaying updating chassis table for 4 seconds run /usr/lib/python3.9/site-packages/neutron/agent/ovn/metadata/agent.py:274 Feb 25 23:59:54 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:59:54.108 28953 INFO ovsdbapp.backend.ovs_idl.vlog [-] ssl:ovsdbserver-sb.openstack.svc:6642: connected Feb 25 23:59:58 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:59:58.100 28859 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): DbSetCommand(_result=None, table=Chassis_Private, record=5cf4be0c-ece4-4570-a909-9a606b819258, col_values=(('external_ids', {'neutron:ovn-metadata-sb-cfg': '1'}),), if_exists=True) do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89 Feb 25 23:59:58 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-25 23:59:58.101 28859 DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Transaction caused no change do_commit /usr/lib/python3.9/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:129 Feb 26 00:00:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:00:27.736 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 00:00:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:00:27.737 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 00:00:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:00:27.737 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 00:01:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:01:27.737 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 00:01:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:01:27.737 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 00:01:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:01:27.737 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 00:02:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:02:27.738 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 00:02:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:02:27.738 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 00:02:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:02:27.738 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 00:03:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:03:27.740 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:404 Feb 26 00:03:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:03:27.740 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:409 Feb 26 00:03:27 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:03:27.740 28859 DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.000s inner /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:423 Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.453 28953 INFO eventlet.wsgi.server [-] (28953) wsgi exited, is_accepting=True Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.454 28953 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.454 28953 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.454 28953 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.457 28859 DEBUG oslo_privsep.comm [-] EOF on privsep read channel _reader_main /usr/lib/python3.9/site-packages/oslo_privsep/comm.py:170 Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.781 28859 INFO oslo_service.service [-] Caught SIGTERM, stopping children Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.782 28859 DEBUG oslo_concurrency.lockutils [-] Acquiring lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:312 Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.783 28859 DEBUG oslo_concurrency.lockutils [-] Acquired lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:315 Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.783 28859 DEBUG oslo_concurrency.lockutils [-] Releasing lock "singleton_lock" lock /usr/lib/python3.9/site-packages/oslo_concurrency/lockutils.py:333 Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.783 28859 DEBUG oslo_service.service [-] Stop services. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:695 Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.784 28859 DEBUG oslo_service.service [-] Killing children. stop /usr/lib/python3.9/site-packages/oslo_service/service.py:700 Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.784 28859 INFO oslo_service.service [-] Waiting on 1 children to exit Feb 26 00:04:08 edpm-compute-0 ovn_metadata_agent[28853]: 2026-02-26 00:04:08.784 28859 INFO oslo_service.service [-] Child 28953 exited with status 0 Feb 26 00:04:12 edpm-compute-0 podman[174974]: Error: no container with ID 917ef442c5f5ce2b1b77f2cb74e43346882e79cfdc931e1bdd3fd50becfa4bbb found in database: no such container Feb 26 00:04:12 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Feb 26 00:04:12 edpm-compute-0 podman[174990]: Error: no container with name or ID "ovn_metadata_agent" found: no such container Feb 26 00:04:12 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Control process exited, code=exited, status=125/n/a Feb 26 00:04:12 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Failed with result 'exit-code'. Feb 26 00:04:12 edpm-compute-0 systemd[1]: edpm_ovn_metadata_agent.service: Scheduled restart job, restart counter is at 1. Feb 26 00:04:12 edpm-compute-0 systemd[1]: Stopped ovn_metadata_agent container. Feb 26 00:04:12 edpm-compute-0 systemd[1]: Starting ovn_metadata_agent container... Feb 26 00:04:13 edpm-compute-0 edpm-start-podman-container[175035]: ovn_metadata_agent Feb 26 00:04:13 edpm-compute-0 edpm-start-podman-container[175032]: Creating additional drop-in dependency for "ovn_metadata_agent" (9b2e91d0321c6a49732f37ef41e2fc39a7b38d73be588d07e1a35f56d83c3cd9) Feb 26 00:04:13 edpm-compute-0 systemd[1]: Started ovn_metadata_agent container.